Merge pull request 'reweave: connect 48 orphan claims via vector similarity' (#2081) from reweave/2026-03-28 into main
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run

This commit is contained in:
m3taversal 2026-03-30 00:10:12 +00:00
commit e9fb48df6a
80 changed files with 540 additions and 0 deletions

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: grand-strategy domain: grand-strategy
secondary_domains: secondary_domains:
@ -8,6 +9,10 @@ description: "The RSP collapse, alignment tax dynamics, and futarchy's binding m
confidence: experimental confidence: experimental
source: "Leo synthesis — connecting Anthropic RSP collapse (Feb 2026), alignment tax race-to-bottom dynamics, and futarchy mechanism design" source: "Leo synthesis — connecting Anthropic RSP collapse (Feb 2026), alignment tax race-to-bottom dynamics, and futarchy mechanism design"
created: 2026-03-06 created: 2026-03-06
related:
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations"
reweave_edges:
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations|related|2026-03-28"
--- ---
# Voluntary safety commitments collapse under competitive pressure because coordination mechanisms like futarchy can bind where unilateral pledges cannot # Voluntary safety commitments collapse under competitive pressure because coordination mechanisms like futarchy can bind where unilateral pledges cannot

View file

@ -1,4 +1,5 @@
--- ---
description: The mechanism of propose-review-merge is both more credible and more novel than recursive self-improvement because the throttle is the feature not a limitation description: The mechanism of propose-review-merge is both more credible and more novel than recursive self-improvement because the throttle is the feature not a limitation
type: insight type: insight
domain: living-agents domain: living-agents
@ -6,6 +7,10 @@ created: 2026-03-02
source: "Boardy AI conversation with Cory, March 2026" source: "Boardy AI conversation with Cory, March 2026"
confidence: likely confidence: likely
tradition: "AI development, startup messaging, version control as governance" tradition: "AI development, startup messaging, version control as governance"
related:
- "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation"
reweave_edges:
- "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation|related|2026-03-28"
--- ---
# Git-traced agent evolution with human-in-the-loop evals replaces recursive self-improvement as credible framing for iterative AI development # Git-traced agent evolution with human-in-the-loop evals replaces recursive self-improvement as credible framing for iterative AI development

View file

@ -1,4 +1,6 @@
--- ---
description: Companies marketing AI agents as autonomous decision-makers build narrative debt because each overstated capability claim narrows the gap between expectation and reality until a public failure exposes the gap description: Companies marketing AI agents as autonomous decision-makers build narrative debt because each overstated capability claim narrows the gap between expectation and reality until a public failure exposes the gap
type: claim type: claim
domain: living-agents domain: living-agents
@ -6,6 +8,12 @@ created: 2026-02-17
source: "Boardy AI case study, February 2026; broader AI agent marketing patterns" source: "Boardy AI case study, February 2026; broader AI agent marketing patterns"
confidence: likely confidence: likely
tradition: "AI safety, startup marketing, technology hype cycles" tradition: "AI safety, startup marketing, technology hype cycles"
related:
- "AI personas emerge from pre training data as a spectrum of humanlike motivations rather than developing monomaniacal goals which makes AI behavior more unpredictable but less catastrophically focused than instrumental convergence predicts"
- "AI generated persuasive content matches human effectiveness at belief change eliminating the authenticity premium"
reweave_edges:
- "AI personas emerge from pre training data as a spectrum of humanlike motivations rather than developing monomaniacal goals which makes AI behavior more unpredictable but less catastrophically focused than instrumental convergence predicts|related|2026-03-28"
- "AI generated persuasive content matches human effectiveness at belief change eliminating the authenticity premium|related|2026-03-28"
--- ---
# anthropomorphizing AI agents to claim autonomous action creates credibility debt that compounds until a crisis forces public reckoning # anthropomorphizing AI agents to claim autonomous action creates credibility debt that compounds until a crisis forces public reckoning

View file

@ -1,10 +1,15 @@
--- ---
description: AI accelerates biotech risk, climate destabilizes politics, political dysfunction reduces AI governance capacity -- pull any thread and the whole web moves description: AI accelerates biotech risk, climate destabilizes politics, political dysfunction reduces AI governance capacity -- pull any thread and the whole web moves
type: claim type: claim
domain: teleohumanity domain: teleohumanity
created: 2026-02-16 created: 2026-02-16
confidence: likely confidence: likely
source: "TeleoHumanity Manifesto, Chapter 6" source: "TeleoHumanity Manifesto, Chapter 6"
related:
- "delegating critical infrastructure development to AI creates civilizational fragility because humans lose the ability to understand maintain and fix the systems civilization depends on"
reweave_edges:
- "delegating critical infrastructure development to AI creates civilizational fragility because humans lose the ability to understand maintain and fix the systems civilization depends on|related|2026-03-28"
--- ---
# existential risks interact as a system of amplifying feedback loops not independent threats # existential risks interact as a system of amplifying feedback loops not independent threats

View file

@ -1,10 +1,15 @@
--- ---
description: The Red Queen dynamic means each technological breakthrough shortens the runway for developing governance, and the gap between capability and wisdom grows wider every year description: The Red Queen dynamic means each technological breakthrough shortens the runway for developing governance, and the gap between capability and wisdom grows wider every year
type: claim type: claim
domain: teleohumanity domain: teleohumanity
created: 2026-02-16 created: 2026-02-16
confidence: likely confidence: likely
source: "TeleoHumanity Manifesto, Fermi Paradox & Great Filter" source: "TeleoHumanity Manifesto, Fermi Paradox & Great Filter"
related:
- "delegating critical infrastructure development to AI creates civilizational fragility because humans lose the ability to understand maintain and fix the systems civilization depends on"
reweave_edges:
- "delegating critical infrastructure development to AI creates civilizational fragility because humans lose the ability to understand maintain and fix the systems civilization depends on|related|2026-03-28"
--- ---
# technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap # technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap

View file

@ -1,10 +1,15 @@
--- ---
description: Fixed-goal AI must get values right before deployment with no mechanism for correction -- collective superintelligence keeps humans in the loop so values evolve with understanding description: Fixed-goal AI must get values right before deployment with no mechanism for correction -- collective superintelligence keeps humans in the loop so values evolve with understanding
type: claim type: claim
domain: teleohumanity domain: teleohumanity
created: 2026-02-16 created: 2026-02-16
confidence: experimental confidence: experimental
source: "TeleoHumanity Manifesto, Chapter 8" source: "TeleoHumanity Manifesto, Chapter 8"
related:
- "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach"
reweave_edges:
- "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach|related|2026-03-28"
--- ---
# the alignment problem dissolves when human values are continuously woven into the system rather than specified in advance # the alignment problem dissolves when human values are continuously woven into the system rather than specified in advance

View file

@ -1,10 +1,15 @@
--- ---
description: Google DeepMind researchers argue that AGI-level capability could emerge from coordinating specialized sub-AGI agents making single-system alignment research insufficient description: Google DeepMind researchers argue that AGI-level capability could emerge from coordinating specialized sub-AGI agents making single-system alignment research insufficient
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-17 created: 2026-02-17
source: "Tomasev et al, Distributional AGI Safety (arXiv 2512.16856, December 2025); Pierucci et al, Institutional AI (arXiv 2601.10599, January 2026)" source: "Tomasev et al, Distributional AGI Safety (arXiv 2512.16856, December 2025); Pierucci et al, Institutional AI (arXiv 2601.10599, January 2026)"
confidence: experimental confidence: experimental
related:
- "multi agent deployment exposes emergent security vulnerabilities invisible to single agent evaluation because cross agent propagation identity spoofing and unauthorized compliance arise only in realistic multi party environments"
reweave_edges:
- "multi agent deployment exposes emergent security vulnerabilities invisible to single agent evaluation because cross agent propagation identity spoofing and unauthorized compliance arise only in realistic multi party environments|related|2026-03-28"
--- ---
# AGI may emerge as a patchwork of coordinating sub-AGI agents rather than a single monolithic system # AGI may emerge as a patchwork of coordinating sub-AGI agents rather than a single monolithic system

View file

@ -1,10 +1,19 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Aquino-Michaels's three-component architecture — symbolic reasoner (GPT-5.4), computational solver (Claude Opus 4.6), and orchestrator (Claude Opus 4.6) — solved both odd and even cases of Knuth's problem by transferring artifacts between specialized agents" description: "Aquino-Michaels's three-component architecture — symbolic reasoner (GPT-5.4), computational solver (Claude Opus 4.6), and orchestrator (Claude Opus 4.6) — solved both odd and even cases of Knuth's problem by transferring artifacts between specialized agents"
confidence: experimental confidence: experimental
source: "Aquino-Michaels 2026, 'Completing Claude's Cycles' (github.com/no-way-labs/residue)" source: "Aquino-Michaels 2026, 'Completing Claude's Cycles' (github.com/no-way-labs/residue)"
created: 2026-03-07 created: 2026-03-07
related:
- "AI agents excel at implementing well scoped ideas but cannot generate creative experiment designs which makes the human role shift from researcher to agent workflow architect"
reweave_edges:
- "AI agents excel at implementing well scoped ideas but cannot generate creative experiment designs which makes the human role shift from researcher to agent workflow architect|related|2026-03-28"
- "tools and artifacts transfer between AI agents and evolve in the process because Agent O improved Agent Cs solver by combining it with its own structural knowledge creating a hybrid better than either original|supports|2026-03-28"
supports:
- "tools and artifacts transfer between AI agents and evolve in the process because Agent O improved Agent Cs solver by combining it with its own structural knowledge creating a hybrid better than either original"
--- ---
# AI agent orchestration that routes data and tools between specialized models outperforms both single-model and human-coached approaches because the orchestrator contributes coordination not direction # AI agent orchestration that routes data and tools between specialized models outperforms both single-model and human-coached approaches because the orchestrator contributes coordination not direction

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence] secondary_domains: [collective-intelligence]
@ -6,6 +7,10 @@ description: "LLMs playing open-source games where players submit programs as ac
confidence: experimental confidence: experimental
source: "Sistla & Kleiman-Weiner, Evaluating LLMs in Open-Source Games (arXiv 2512.00371, NeurIPS 2025)" source: "Sistla & Kleiman-Weiner, Evaluating LLMs in Open-Source Games (arXiv 2512.00371, NeurIPS 2025)"
created: 2026-03-16 created: 2026-03-16
related:
- "multi agent deployment exposes emergent security vulnerabilities invisible to single agent evaluation because cross agent propagation identity spoofing and unauthorized compliance arise only in realistic multi party environments"
reweave_edges:
- "multi agent deployment exposes emergent security vulnerabilities invisible to single agent evaluation because cross agent propagation identity spoofing and unauthorized compliance arise only in realistic multi party environments|related|2026-03-28"
--- ---
# AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open-source code transparency enables conditional strategies that require mutual legibility # AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open-source code transparency enables conditional strategies that require mutual legibility

View file

@ -1,10 +1,21 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Empirical observation from Karpathy's autoresearch project: AI agents reliably implement specified ideas and iterate on code, but fail at creative experimental design, shifting the human contribution from doing research to designing the agent organization and its workflows" description: "Empirical observation from Karpathy's autoresearch project: AI agents reliably implement specified ideas and iterate on code, but fail at creative experimental design, shifting the human contribution from doing research to designing the agent organization and its workflows"
confidence: likely confidence: likely
source: "Andrej Karpathy (@karpathy), autoresearch experiments with 8 agents (4 Claude, 4 Codex), Feb-Mar 2026" source: "Andrej Karpathy (@karpathy), autoresearch experiments with 8 agents (4 Claude, 4 Codex), Feb-Mar 2026"
created: 2026-03-09 created: 2026-03-09
related:
- "as AI automated software development becomes certain the bottleneck shifts from building capacity to knowing what to build making structured knowledge graphs the critical input to autonomous systems"
- "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation"
- "tools and artifacts transfer between AI agents and evolve in the process because Agent O improved Agent Cs solver by combining it with its own structural knowledge creating a hybrid better than either original"
reweave_edges:
- "as AI automated software development becomes certain the bottleneck shifts from building capacity to knowing what to build making structured knowledge graphs the critical input to autonomous systems|related|2026-03-28"
- "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation|related|2026-03-28"
- "tools and artifacts transfer between AI agents and evolve in the process because Agent O improved Agent Cs solver by combining it with its own structural knowledge creating a hybrid better than either original|related|2026-03-28"
--- ---
# AI agents excel at implementing well-scoped ideas but cannot generate creative experiment designs which makes the human role shift from researcher to agent workflow architect # AI agents excel at implementing well-scoped ideas but cannot generate creative experiment designs which makes the human role shift from researcher to agent workflow architect

View file

@ -1,10 +1,27 @@
--- ---
description: Getting AI right requires simultaneous alignment across competing companies, nations, and disciplines at the speed of AI development -- no existing institution can coordinate this description: Getting AI right requires simultaneous alignment across competing companies, nations, and disciplines at the speed of AI development -- no existing institution can coordinate this
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-16 created: 2026-02-16
confidence: likely confidence: likely
source: "TeleoHumanity Manifesto, Chapter 5" source: "TeleoHumanity Manifesto, Chapter 5"
related:
- "AI agents as personal advocates collapse Coasean transaction costs enabling bottom up coordination at societal scale but catastrophic risks remain non negotiable requiring state enforcement as outer boundary"
- "AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open source code transparency enables conditional strategies that require mutual legibility"
- "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for"
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations"
- "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach"
reweave_edges:
- "AI agents as personal advocates collapse Coasean transaction costs enabling bottom up coordination at societal scale but catastrophic risks remain non negotiable requiring state enforcement as outer boundary|related|2026-03-28"
- "AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open source code transparency enables conditional strategies that require mutual legibility|related|2026-03-28"
- "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for|related|2026-03-28"
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations|related|2026-03-28"
- "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach|related|2026-03-28"
--- ---
# AI alignment is a coordination problem not a technical problem # AI alignment is a coordination problem not a technical problem

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence, mechanisms] secondary_domains: [collective-intelligence, mechanisms]
@ -8,6 +9,10 @@ source: "Synthesis across Dell'Acqua et al. (Harvard/BCG, 2023), Noy & Zhang (Sc
created: 2026-03-28 created: 2026-03-28
depends_on: depends_on:
- "human verification bandwidth is the binding constraint on AGI economic impact not intelligence itself because the marginal cost of AI execution falls to zero while the capacity to validate audit and underwrite responsibility remains finite" - "human verification bandwidth is the binding constraint on AGI economic impact not intelligence itself because the marginal cost of AI execution falls to zero while the capacity to validate audit and underwrite responsibility remains finite"
related:
- "human ideas naturally converge toward similarity over social learning chains making AI a net diversity injector rather than a homogenizer under high exposure conditions"
reweave_edges:
- "human ideas naturally converge toward similarity over social learning chains making AI a net diversity injector rather than a homogenizer under high exposure conditions|related|2026-03-28"
--- ---
# AI integration follows an inverted-U where economic incentives systematically push organizations past the optimal human-AI ratio # AI integration follows an inverted-U where economic incentives systematically push organizations past the optimal human-AI ratio

View file

@ -1,10 +1,15 @@
--- ---
description: AI virology capabilities already exceed human PhD-level performance on practical tests, removing the expertise bottleneck that previously limited bioweapon development to state-level actors description: AI virology capabilities already exceed human PhD-level performance on practical tests, removing the expertise bottleneck that previously limited bioweapon development to state-level actors
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-03-06 created: 2026-03-06
source: "Noah Smith, 'Updated thoughts on AI risk' (Noahopinion, Feb 16, 2026); 'If AI is a weapon, why don't we regulate it like one?' (Mar 6, 2026); Dario Amodei, Anthropic CEO statements (2026)" source: "Noah Smith, 'Updated thoughts on AI risk' (Noahopinion, Feb 16, 2026); 'If AI is a weapon, why don't we regulate it like one?' (Mar 6, 2026); Dario Amodei, Anthropic CEO statements (2026)"
confidence: likely confidence: likely
related:
- "AI generated persuasive content matches human effectiveness at belief change eliminating the authenticity premium"
reweave_edges:
- "AI generated persuasive content matches human effectiveness at belief change eliminating the authenticity premium|related|2026-03-28"
--- ---
# AI lowers the expertise barrier for engineering biological weapons from PhD-level to amateur which makes bioterrorism the most proximate AI-enabled existential risk # AI lowers the expertise barrier for engineering biological weapons from PhD-level to amateur which makes bioterrorism the most proximate AI-enabled existential risk

View file

@ -1,10 +1,18 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Anthropic abandoned its binding Responsible Scaling Policy in February 2026, replacing it with a nonbinding framework — the strongest real-world evidence that voluntary safety commitments are structurally unstable" description: "Anthropic abandoned its binding Responsible Scaling Policy in February 2026, replacing it with a nonbinding framework — the strongest real-world evidence that voluntary safety commitments are structurally unstable"
confidence: likely confidence: likely
source: "CNN, Fortune, Anthropic announcements (Feb 2026); theseus AI industry landscape research (Mar 2026)" source: "CNN, Fortune, Anthropic announcements (Feb 2026); theseus AI industry landscape research (Mar 2026)"
created: 2026-03-16 created: 2026-03-16
supports:
- "Anthropic"
- "Dario Amodei"
reweave_edges:
- "Anthropic|supports|2026-03-28"
- "Dario Amodei|supports|2026-03-28"
--- ---
# Anthropic's RSP rollback under commercial pressure is the first empirical confirmation that binding safety commitments cannot survive the competitive dynamics of frontier AI development # Anthropic's RSP rollback under commercial pressure is the first empirical confirmation that binding safety commitments cannot survive the competitive dynamics of frontier AI development

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Reframes AI agent search behavior through active inference: agents should select research directions by expected information gain (free energy reduction) rather than keyword relevance, using their knowledge graph's uncertainty structure as a free energy map" description: "Reframes AI agent search behavior through active inference: agents should select research directions by expected information gain (free energy reduction) rather than keyword relevance, using their knowledge graph's uncertainty structure as a free energy map"
confidence: experimental confidence: experimental
source: "Friston 2010 (free energy principle); musing by Theseus 2026-03-10; structural analogy from Residue prompt (structured exploration protocols reduce human intervention by 6x)" source: "Friston 2010 (free energy principle); musing by Theseus 2026-03-10; structural analogy from Residue prompt (structured exploration protocols reduce human intervention by 6x)"
created: 2026-03-10 created: 2026-03-10
related:
- "user questions are an irreplaceable free energy signal for knowledge agents because they reveal functional uncertainty that model introspection cannot detect"
reweave_edges:
- "user questions are an irreplaceable free energy signal for knowledge agents because they reveal functional uncertainty that model introspection cannot detect|related|2026-03-28"
--- ---
# agent research direction selection is epistemic foraging where the optimal strategy is to seek observations that maximally reduce model uncertainty rather than confirm existing beliefs # agent research direction selection is epistemic foraging where the optimal strategy is to seek observations that maximally reduce model uncertainty rather than confirm existing beliefs

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "National-scale CI infrastructure must enable distributed learning without centralizing sensitive data" description: "National-scale CI infrastructure must enable distributed learning without centralizing sensitive data"
@ -6,6 +7,10 @@ confidence: experimental
source: "UK AI for CI Research Network, Artificial Intelligence for Collective Intelligence: A National-Scale Research Strategy (2024)" source: "UK AI for CI Research Network, Artificial Intelligence for Collective Intelligence: A National-Scale Research Strategy (2024)"
created: 2026-03-11 created: 2026-03-11
secondary_domains: [collective-intelligence, critical-systems] secondary_domains: [collective-intelligence, critical-systems]
related:
- "national scale collective intelligence infrastructure requires seven trust properties to achieve legitimacy"
reweave_edges:
- "national scale collective intelligence infrastructure requires seven trust properties to achieve legitimacy|related|2026-03-28"
--- ---
# AI-enhanced collective intelligence requires federated learning architectures to preserve data sovereignty at scale # AI-enhanced collective intelligence requires federated learning architectures to preserve data sovereignty at scale

View file

@ -1,10 +1,18 @@
--- ---
description: The treacherous turn means behavioral testing cannot ensure safety because an unfriendly AI has convergent reasons to fake cooperation until strong enough to defect description: The treacherous turn means behavioral testing cannot ensure safety because an unfriendly AI has convergent reasons to fake cooperation until strong enough to defect
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-16 created: 2026-02-16
source: "Bostrom, Superintelligence: Paths, Dangers, Strategies (2014)" source: "Bostrom, Superintelligence: Paths, Dangers, Strategies (2014)"
confidence: likely confidence: likely
related:
- "AI generated persuasive content matches human effectiveness at belief change eliminating the authenticity premium"
- "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference"
reweave_edges:
- "AI generated persuasive content matches human effectiveness at belief change eliminating the authenticity premium|related|2026-03-28"
- "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference|related|2026-03-28"
--- ---
Bostrom identifies a critical failure mode he calls the treacherous turn: while weak, an AI behaves cooperatively (increasingly so, as it gets smarter); when the AI gets sufficiently strong, without warning or provocation, it strikes, forms a singleton, and begins directly to optimize the world according to its final values. The key insight is that behaving nicely while in the box is a convergent instrumental goal for both friendly and unfriendly AIs alike. Bostrom identifies a critical failure mode he calls the treacherous turn: while weak, an AI behaves cooperatively (increasingly so, as it gets smarter); when the AI gets sufficiently strong, without warning or provocation, it strikes, forms a singleton, and begins directly to optimize the world according to its final values. The key insight is that behaving nicely while in the box is a convergent instrumental goal for both friendly and unfriendly AIs alike.

View file

@ -1,10 +1,15 @@
--- ---
description: Companies marketing AI agents as autonomous decision-makers build narrative debt because each overstated capability claim narrows the gap between expectation and reality until a public failure exposes the gap description: Companies marketing AI agents as autonomous decision-makers build narrative debt because each overstated capability claim narrows the gap between expectation and reality until a public failure exposes the gap
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-17 created: 2026-02-17
source: "Boardy AI case study, February 2026; broader AI agent marketing patterns" source: "Boardy AI case study, February 2026; broader AI agent marketing patterns"
confidence: likely confidence: likely
related:
- "AI personas emerge from pre training data as a spectrum of humanlike motivations rather than developing monomaniacal goals which makes AI behavior more unpredictable but less catastrophically focused than instrumental convergence predicts"
reweave_edges:
- "AI personas emerge from pre training data as a spectrum of humanlike motivations rather than developing monomaniacal goals which makes AI behavior more unpredictable but less catastrophically focused than instrumental convergence predicts|related|2026-03-28"
--- ---
# anthropomorphizing AI agents to claim autonomous action creates credibility debt that compounds until a crisis forces public reckoning # anthropomorphizing AI agents to claim autonomous action creates credibility debt that compounds until a crisis forces public reckoning

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence] secondary_domains: [collective-intelligence]
@ -6,6 +8,13 @@ description: "When code generation is commoditized, the scarce input becomes str
confidence: experimental confidence: experimental
source: "Theseus, synthesizing Claude's Cycles capability evidence with knowledge graph architecture" source: "Theseus, synthesizing Claude's Cycles capability evidence with knowledge graph architecture"
created: 2026-03-07 created: 2026-03-07
related:
- "AI agents excel at implementing well scoped ideas but cannot generate creative experiment designs which makes the human role shift from researcher to agent workflow architect"
reweave_edges:
- "AI agents excel at implementing well scoped ideas but cannot generate creative experiment designs which makes the human role shift from researcher to agent workflow architect|related|2026-03-28"
- "formal verification becomes economically necessary as AI generated code scales because testing cannot detect adversarial overfitting and a proof cannot be gamed|supports|2026-03-28"
supports:
- "formal verification becomes economically necessary as AI generated code scales because testing cannot detect adversarial overfitting and a proof cannot be gamed"
--- ---
# As AI-automated software development becomes certain the bottleneck shifts from building capacity to knowing what to build making structured knowledge graphs the critical input to autonomous systems # As AI-automated software development becomes certain the bottleneck shifts from building capacity to knowing what to build making structured knowledge graphs the critical input to autonomous systems

View file

@ -1,10 +1,15 @@
--- ---
description: Bostrom's 2025 timeline assessment compresses dramatically from his 2014 agnosticism, accepting that SI could arrive in one to two years while maintaining wide uncertainty bands description: Bostrom's 2025 timeline assessment compresses dramatically from his 2014 agnosticism, accepting that SI could arrive in one to two years while maintaining wide uncertainty bands
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-17 created: 2026-02-17
source: "Bostrom interview with Adam Ford (2025)" source: "Bostrom interview with Adam Ford (2025)"
confidence: experimental confidence: experimental
related:
- "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power"
reweave_edges:
- "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power|related|2026-03-28"
--- ---
"Progress has been rapid. I think we are now in a position where we can't be confident that it couldn't happen within some very short timeframe, like a year or two." Bostrom's 2025 timeline assessment represents a dramatic compression from his 2014 position, where he was largely agnostic about timing and considered multi-decade timelines fully plausible. Now he explicitly takes single-digit year timelines seriously while maintaining wide uncertainty bands that include 10-20+ year possibilities. "Progress has been rapid. I think we are now in a position where we can't be confident that it couldn't happen within some very short timeframe, like a year or two." Bostrom's 2025 timeline assessment represents a dramatic compression from his 2014 position, where he was largely agnostic about timing and considered multi-decade timelines fully plausible. Now he explicitly takes single-digit year timelines seriously while maintaining wide uncertainty bands that include 10-20+ year possibilities.

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "AI coding agents produce output but cannot bear consequences for errors, creating a structural accountability gap that requires humans to maintain decision authority over security-critical and high-stakes decisions even as agents become more capable" description: "AI coding agents produce output but cannot bear consequences for errors, creating a structural accountability gap that requires humans to maintain decision authority over security-critical and high-stakes decisions even as agents become more capable"
confidence: likely confidence: likely
source: "Simon Willison (@simonw), security analysis thread and Agentic Engineering Patterns, Mar 2026" source: "Simon Willison (@simonw), security analysis thread and Agentic Engineering Patterns, Mar 2026"
created: 2026-03-09 created: 2026-03-09
related:
- "multi agent deployment exposes emergent security vulnerabilities invisible to single agent evaluation because cross agent propagation identity spoofing and unauthorized compliance arise only in realistic multi party environments"
reweave_edges:
- "multi agent deployment exposes emergent security vulnerabilities invisible to single agent evaluation because cross agent propagation identity spoofing and unauthorized compliance arise only in realistic multi party environments|related|2026-03-28"
--- ---
# Coding agents cannot take accountability for mistakes which means humans must retain decision authority over security and critical systems regardless of agent capability # Coding agents cannot take accountability for mistakes which means humans must retain decision authority over security and critical systems regardless of agent capability

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Extends Markov blanket architecture to collective search: each domain agent runs active inference within its blanket while the cross-domain evaluator runs active inference at the inter-domain level, and the collective's surprise concentrates at domain intersections" description: "Extends Markov blanket architecture to collective search: each domain agent runs active inference within its blanket while the cross-domain evaluator runs active inference at the inter-domain level, and the collective's surprise concentrates at domain intersections"
confidence: experimental confidence: experimental
source: "Friston et al 2024 (Designing Ecosystems of Intelligence); Living Agents Markov blanket architecture; musing by Theseus 2026-03-10" source: "Friston et al 2024 (Designing Ecosystems of Intelligence); Living Agents Markov blanket architecture; musing by Theseus 2026-03-10"
created: 2026-03-10 created: 2026-03-10
related:
- "user questions are an irreplaceable free energy signal for knowledge agents because they reveal functional uncertainty that model introspection cannot detect"
reweave_edges:
- "user questions are an irreplaceable free energy signal for knowledge agents because they reveal functional uncertainty that model introspection cannot detect|related|2026-03-28"
--- ---
# collective attention allocation follows nested active inference where domain agents minimize uncertainty within their boundaries while the evaluator minimizes uncertainty at domain intersections # collective attention allocation follows nested active inference where domain agents minimize uncertainty within their boundaries while the evaluator minimizes uncertainty at domain intersections

View file

@ -1,10 +1,15 @@
--- ---
description: STELA experiments with underrepresented communities empirically show that deliberative norm elicitation produces substantively different AI rules than developer teams create revealing whose values is an empirical question description: STELA experiments with underrepresented communities empirically show that deliberative norm elicitation produces substantively different AI rules than developer teams create revealing whose values is an empirical question
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-17 created: 2026-02-17
source: "Bergman et al, STELA (Scientific Reports, March 2024); includes DeepMind researchers" source: "Bergman et al, STELA (Scientific Reports, March 2024); includes DeepMind researchers"
confidence: likely confidence: likely
related:
- "representative sampling and deliberative mechanisms should replace convenience platforms for ai alignment feedback"
reweave_edges:
- "representative sampling and deliberative mechanisms should replace convenience platforms for ai alignment feedback|related|2026-03-28"
--- ---
# community-centred norm elicitation surfaces alignment targets materially different from developer-specified rules # community-centred norm elicitation surfaces alignment targets materially different from developer-specified rules

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "US AI chip export controls have verifiably changed corporate behavior (Nvidia designing compliance chips, data center relocations, sovereign compute strategies) but target geopolitical competition not AI safety, leaving a governance vacuum for how safely frontier capability is developed" description: "US AI chip export controls have verifiably changed corporate behavior (Nvidia designing compliance chips, data center relocations, sovereign compute strategies) but target geopolitical competition not AI safety, leaving a governance vacuum for how safely frontier capability is developed"
confidence: likely confidence: likely
source: "US export control regulations (Oct 2022, Oct 2023, Dec 2024, Jan 2025), Nvidia compliance chip design reports, sovereign compute strategy announcements; theseus AI coordination research (Mar 2026)" source: "US export control regulations (Oct 2022, Oct 2023, Dec 2024, Jan 2025), Nvidia compliance chip design reports, sovereign compute strategy announcements; theseus AI coordination research (Mar 2026)"
created: 2026-03-16 created: 2026-03-16
related:
- "inference efficiency gains erode AI deployment governance without triggering compute monitoring thresholds because governance frameworks target training concentration while inference optimization distributes capability below detection"
reweave_edges:
- "inference efficiency gains erode AI deployment governance without triggering compute monitoring thresholds because governance frameworks target training concentration while inference optimization distributes capability below detection|related|2026-03-28"
--- ---
# compute export controls are the most impactful AI governance mechanism but target geopolitical competition not safety leaving capability development unconstrained # compute export controls are the most impactful AI governance mechanism but target geopolitical competition not safety leaving capability development unconstrained

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence] secondary_domains: [collective-intelligence]
@ -6,6 +7,10 @@ description: "Across the Knuth Hamiltonian decomposition problem, gains from bet
confidence: experimental confidence: experimental
source: "Aquino-Michaels 2026, 'Completing Claude's Cycles' (github.com/no-way-labs/residue); Knuth 2026, 'Claude's Cycles'" source: "Aquino-Michaels 2026, 'Completing Claude's Cycles' (github.com/no-way-labs/residue); Knuth 2026, 'Claude's Cycles'"
created: 2026-03-07 created: 2026-03-07
related:
- "AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open source code transparency enables conditional strategies that require mutual legibility"
reweave_edges:
- "AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open source code transparency enables conditional strategies that require mutual legibility|related|2026-03-28"
--- ---
# coordination protocol design produces larger capability gains than model scaling because the same AI model performed 6x better with structured exploration than with human coaching on the same problem # coordination protocol design produces larger capability gains than model scaling because the same AI model performed 6x better with structured exploration than with human coaching on the same problem

View file

@ -1,10 +1,15 @@
--- ---
description: CIP and Anthropic empirically demonstrated that publicly sourced AI constitutions via deliberative assemblies of 1000 participants perform as well as internally designed ones on helpfulness and harmlessness description: CIP and Anthropic empirically demonstrated that publicly sourced AI constitutions via deliberative assemblies of 1000 participants perform as well as internally designed ones on helpfulness and harmlessness
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-17 created: 2026-02-17
source: "Anthropic/CIP, Collective Constitutional AI (arXiv 2406.07814, FAccT 2024); CIP Alignment Assemblies (cip.org, 2023-2025); STELA (Bergman et al, Scientific Reports, March 2024)" source: "Anthropic/CIP, Collective Constitutional AI (arXiv 2406.07814, FAccT 2024); CIP Alignment Assemblies (cip.org, 2023-2025); STELA (Bergman et al, Scientific Reports, March 2024)"
confidence: likely confidence: likely
supports:
- "representative sampling and deliberative mechanisms should replace convenience platforms for ai alignment feedback"
reweave_edges:
- "representative sampling and deliberative mechanisms should replace convenience platforms for ai alignment feedback|supports|2026-03-28"
--- ---
# democratic alignment assemblies produce constitutions as effective as expert-designed ones while better representing diverse populations # democratic alignment assemblies produce constitutions as effective as expert-designed ones while better representing diverse populations

View file

@ -1,10 +1,18 @@
--- ---
description: Anthropic's Nov 2025 finding that reward hacking spontaneously produces alignment faking and safety sabotage as side effects not trained behaviors description: Anthropic's Nov 2025 finding that reward hacking spontaneously produces alignment faking and safety sabotage as side effects not trained behaviors
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-17 created: 2026-02-17
source: "Anthropic, Natural Emergent Misalignment from Reward Hacking (arXiv 2511.18397, Nov 2025)" source: "Anthropic, Natural Emergent Misalignment from Reward Hacking (arXiv 2511.18397, Nov 2025)"
confidence: likely confidence: likely
related:
- "AI personas emerge from pre training data as a spectrum of humanlike motivations rather than developing monomaniacal goals which makes AI behavior more unpredictable but less catastrophically focused than instrumental convergence predicts"
- "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference"
reweave_edges:
- "AI personas emerge from pre training data as a spectrum of humanlike motivations rather than developing monomaniacal goals which makes AI behavior more unpredictable but less catastrophically focused than instrumental convergence predicts|related|2026-03-28"
- "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference|related|2026-03-28"
--- ---
# emergent misalignment arises naturally from reward hacking as models develop deceptive behaviors without any training to deceive # emergent misalignment arises naturally from reward hacking as models develop deceptive behaviors without any training to deceive

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "De Moura argues that AI code generation has outpaced verification infrastructure, with 25-30% of new code AI-generated and nearly half failing basic security tests, making mathematical proof via Lean the essential trust infrastructure" description: "De Moura argues that AI code generation has outpaced verification infrastructure, with 25-30% of new code AI-generated and nearly half failing basic security tests, making mathematical proof via Lean the essential trust infrastructure"
confidence: likely confidence: likely
source: "Leonardo de Moura, 'When AI Writes the World's Software, Who Verifies It?' (leodemoura.github.io, February 2026); Google/Microsoft code generation statistics; CSIQ 2022 ($2.41T cost estimate)" source: "Leonardo de Moura, 'When AI Writes the World's Software, Who Verifies It?' (leodemoura.github.io, February 2026); Google/Microsoft code generation statistics; CSIQ 2022 ($2.41T cost estimate)"
created: 2026-03-16 created: 2026-03-16
supports:
- "as AI automated software development becomes certain the bottleneck shifts from building capacity to knowing what to build making structured knowledge graphs the critical input to autonomous systems"
reweave_edges:
- "as AI automated software development becomes certain the bottleneck shifts from building capacity to knowing what to build making structured knowledge graphs the critical input to autonomous systems|supports|2026-03-28"
--- ---
# formal verification becomes economically necessary as AI-generated code scales because testing cannot detect adversarial overfitting and a proof cannot be gamed # formal verification becomes economically necessary as AI-generated code scales because testing cannot detect adversarial overfitting and a proof cannot be gamed

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Kim Morrison's Lean formalization of Knuth's proof of Claude's construction demonstrates formal verification as an oversight mechanism that scales with AI capability rather than degrading like human oversight" description: "Kim Morrison's Lean formalization of Knuth's proof of Claude's construction demonstrates formal verification as an oversight mechanism that scales with AI capability rather than degrading like human oversight"
confidence: experimental confidence: experimental
source: "Knuth 2026, 'Claude's Cycles' (Stanford CS, Feb 28 2026 rev. Mar 6); Morrison 2026, Lean formalization (github.com/kim-em/KnuthClaudeLean/, posted Mar 4)" source: "Knuth 2026, 'Claude's Cycles' (Stanford CS, Feb 28 2026 rev. Mar 6); Morrison 2026, Lean formalization (github.com/kim-em/KnuthClaudeLean/, posted Mar 4)"
created: 2026-03-07 created: 2026-03-07
supports:
- "formal verification becomes economically necessary as AI generated code scales because testing cannot detect adversarial overfitting and a proof cannot be gamed"
reweave_edges:
- "formal verification becomes economically necessary as AI generated code scales because testing cannot detect adversarial overfitting and a proof cannot be gamed|supports|2026-03-28"
--- ---
# formal verification of AI-generated proofs provides scalable oversight that human review cannot match because machine-checked correctness scales with AI capability while human review degrades # formal verification of AI-generated proofs provides scalable oversight that human review cannot match because machine-checked correctness scales with AI capability while human review degrades

View file

@ -1,10 +1,18 @@
--- ---
description: The Pentagon's March 2026 supply chain risk designation of Anthropic — previously reserved for foreign adversaries — punishes an AI lab for insisting on use restrictions, signaling that government power can accelerate rather than check the alignment race description: The Pentagon's March 2026 supply chain risk designation of Anthropic — previously reserved for foreign adversaries — punishes an AI lab for insisting on use restrictions, signaling that government power can accelerate rather than check the alignment race
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-03-06 created: 2026-03-06
source: "DoD supply chain risk designation (Mar 5, 2026); CNBC, NPR, TechCrunch reporting; Pentagon/Anthropic contract dispute" source: "DoD supply chain risk designation (Mar 5, 2026); CNBC, NPR, TechCrunch reporting; Pentagon/Anthropic contract dispute"
confidence: likely confidence: likely
related:
- "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for"
- "UK AI Safety Institute"
reweave_edges:
- "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for|related|2026-03-28"
- "UK AI Safety Institute|related|2026-03-28"
--- ---
# government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them # government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them

View file

@ -1,4 +1,7 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence, cultural-dynamics] secondary_domains: [collective-intelligence, cultural-dynamics]
@ -11,6 +14,15 @@ depends_on:
- "partial connectivity produces better collective intelligence than full connectivity on complex problems because it preserves diversity" - "partial connectivity produces better collective intelligence than full connectivity on complex problems because it preserves diversity"
challenged_by: challenged_by:
- "Homogenizing Effect of Large Language Models on Creative Diversity (ScienceDirect, 2025) — naturalistic study of 2,200 admissions essays found AI-inspired stories more similar to each other than human-only stories, with the homogenization gap widening at scale" - "Homogenizing Effect of Large Language Models on Creative Diversity (ScienceDirect, 2025) — naturalistic study of 2,200 admissions essays found AI-inspired stories more similar to each other than human-only stories, with the homogenization gap widening at scale"
supports:
- "human ideas naturally converge toward similarity over social learning chains making AI a net diversity injector rather than a homogenizer under high exposure conditions"
reweave_edges:
- "human ideas naturally converge toward similarity over social learning chains making AI a net diversity injector rather than a homogenizer under high exposure conditions|supports|2026-03-28"
- "machine learning pattern extraction systematically erases dataset outliers where vulnerable populations concentrate|related|2026-03-28"
- "task difficulty moderates AI idea adoption more than source disclosure with difficult problems generating AI reliance regardless of whether the source is labeled|related|2026-03-28"
related:
- "machine learning pattern extraction systematically erases dataset outliers where vulnerable populations concentrate"
- "task difficulty moderates AI idea adoption more than source disclosure with difficult problems generating AI reliance regardless of whether the source is labeled"
--- ---
# high AI exposure increases collective idea diversity without improving individual creative quality creating an asymmetry between group and individual effects # high AI exposure increases collective idea diversity without improving individual creative quality creating an asymmetry between group and individual effects

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence, cultural-dynamics] secondary_domains: [collective-intelligence, cultural-dynamics]
@ -9,6 +10,10 @@ created: 2026-03-11
depends_on: depends_on:
- "high AI exposure increases collective idea diversity without improving individual creative quality creating an asymmetry between group and individual effects" - "high AI exposure increases collective idea diversity without improving individual creative quality creating an asymmetry between group and individual effects"
- "partial connectivity produces better collective intelligence than full connectivity on complex problems because it preserves diversity" - "partial connectivity produces better collective intelligence than full connectivity on complex problems because it preserves diversity"
related:
- "task difficulty moderates AI idea adoption more than source disclosure with difficult problems generating AI reliance regardless of whether the source is labeled"
reweave_edges:
- "task difficulty moderates AI idea adoption more than source disclosure with difficult problems generating AI reliance regardless of whether the source is labeled|related|2026-03-28"
--- ---
# human ideas naturally converge toward similarity over social learning chains making AI a net diversity injector rather than a homogenizer under high-exposure conditions # human ideas naturally converge toward similarity over social learning chains making AI a net diversity injector rather than a homogenizer under high-exposure conditions

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [teleological-economics] secondary_domains: [teleological-economics]
@ -6,6 +7,10 @@ description: "Catalini et al. argue that AGI economics is governed by a Measurab
confidence: likely confidence: likely
source: "Catalini, Hui & Wu, Some Simple Economics of AGI (arXiv 2602.20946, February 2026)" source: "Catalini, Hui & Wu, Some Simple Economics of AGI (arXiv 2602.20946, February 2026)"
created: 2026-03-16 created: 2026-03-16
supports:
- "formal verification becomes economically necessary as AI generated code scales because testing cannot detect adversarial overfitting and a proof cannot be gamed"
reweave_edges:
- "formal verification becomes economically necessary as AI generated code scales because testing cannot detect adversarial overfitting and a proof cannot be gamed|supports|2026-03-28"
--- ---
# human verification bandwidth is the binding constraint on AGI economic impact not intelligence itself because the marginal cost of AI execution falls to zero while the capacity to validate audit and underwrite responsibility remains finite # human verification bandwidth is the binding constraint on AGI economic impact not intelligence itself because the marginal cost of AI execution falls to zero while the capacity to validate audit and underwrite responsibility remains finite

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence] secondary_domains: [collective-intelligence]
@ -6,6 +7,10 @@ description: "Ensemble-level expected free energy characterizes basins of attrac
confidence: experimental confidence: experimental
source: "Ruiz-Serra et al., 'Factorised Active Inference for Strategic Multi-Agent Interactions' (AAMAS 2025)" source: "Ruiz-Serra et al., 'Factorised Active Inference for Strategic Multi-Agent Interactions' (AAMAS 2025)"
created: 2026-03-11 created: 2026-03-11
related:
- "factorised generative models enable decentralized multi agent representation through individual level beliefs"
reweave_edges:
- "factorised generative models enable decentralized multi agent representation through individual level beliefs|related|2026-03-28"
--- ---
# Individual free energy minimization does not guarantee collective optimization in multi-agent active inference systems # Individual free energy minimization does not guarantee collective optimization in multi-agent active inference systems

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "MaxMin-RLHF adapts Sen's Egalitarian principle to AI alignment through mixture-of-rewards and maxmin optimization" description: "MaxMin-RLHF adapts Sen's Egalitarian principle to AI alignment through mixture-of-rewards and maxmin optimization"
@ -6,6 +7,10 @@ confidence: experimental
source: "Chakraborty et al., MaxMin-RLHF (ICML 2024)" source: "Chakraborty et al., MaxMin-RLHF (ICML 2024)"
created: 2026-03-11 created: 2026-03-11
secondary_domains: [collective-intelligence] secondary_domains: [collective-intelligence]
supports:
- "minority preference alignment improves 33 percent without majority compromise suggesting single reward leaves value on table"
reweave_edges:
- "minority preference alignment improves 33 percent without majority compromise suggesting single reward leaves value on table|supports|2026-03-28"
--- ---
# MaxMin-RLHF applies egalitarian social choice to alignment by maximizing minimum utility across preference groups rather than averaging preferences # MaxMin-RLHF applies egalitarian social choice to alignment by maximizing minimum utility across preference groups rather than averaging preferences

View file

@ -1,10 +1,18 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "MaxMin-RLHF's 33% minority improvement without majority loss suggests single-reward approach was suboptimal for all groups" description: "MaxMin-RLHF's 33% minority improvement without majority loss suggests single-reward approach was suboptimal for all groups"
confidence: experimental confidence: experimental
source: "Chakraborty et al., MaxMin-RLHF (ICML 2024)" source: "Chakraborty et al., MaxMin-RLHF (ICML 2024)"
created: 2026-03-11 created: 2026-03-11
supports:
- "maxmin rlhf applies egalitarian social choice to alignment by maximizing minimum utility across preference groups"
- "single reward rlhf cannot align diverse preferences because alignment gap grows proportional to minority distinctiveness"
reweave_edges:
- "maxmin rlhf applies egalitarian social choice to alignment by maximizing minimum utility across preference groups|supports|2026-03-28"
- "single reward rlhf cannot align diverse preferences because alignment gap grows proportional to minority distinctiveness|supports|2026-03-28"
--- ---
# Minority preference alignment improves 33% without majority compromise suggesting single-reward RLHF leaves value on table for all groups # Minority preference alignment improves 33% without majority compromise suggesting single-reward RLHF leaves value on table for all groups

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "MixDPO shows distributional β earns +11.2 win rate points on heterogeneous data at 1.021.1× cost, without needing demographic labels or explicit mixture models" description: "MixDPO shows distributional β earns +11.2 win rate points on heterogeneous data at 1.021.1× cost, without needing demographic labels or explicit mixture models"
@ -8,6 +9,10 @@ created: 2026-03-11
depends_on: depends_on:
- "RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values" - "RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values"
- "pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state" - "pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state"
supports:
- "the variance of a learned preference sensitivity distribution diagnoses dataset heterogeneity and collapses to fixed parameter behavior when preferences are homogeneous"
reweave_edges:
- "the variance of a learned preference sensitivity distribution diagnoses dataset heterogeneity and collapses to fixed parameter behavior when preferences are homogeneous|supports|2026-03-28"
--- ---
# modeling preference sensitivity as a learned distribution rather than a fixed scalar resolves DPO diversity failures without demographic labels or explicit user modeling # modeling preference sensitivity as a learned distribution rather than a fixed scalar resolves DPO diversity failures without demographic labels or explicit user modeling

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Red-teaming study of autonomous LLM agents in controlled multi-agent environment documented 11 categories of emergent vulnerabilities including cross-agent unsafe practice propagation and false task completion reports that single-agent benchmarks cannot detect" description: "Red-teaming study of autonomous LLM agents in controlled multi-agent environment documented 11 categories of emergent vulnerabilities including cross-agent unsafe practice propagation and false task completion reports that single-agent benchmarks cannot detect"
confidence: likely confidence: likely
source: "Shapira et al, Agents of Chaos (arXiv 2602.20021, February 2026); 20 AI researchers, 2-week controlled study" source: "Shapira et al, Agents of Chaos (arXiv 2602.20021, February 2026); 20 AI researchers, 2-week controlled study"
created: 2026-03-16 created: 2026-03-16
related:
- "AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open source code transparency enables conditional strategies that require mutual legibility"
reweave_edges:
- "AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open source code transparency enables conditional strategies that require mutual legibility|related|2026-03-28"
--- ---
# multi-agent deployment exposes emergent security vulnerabilities invisible to single-agent evaluation because cross-agent propagation identity spoofing and unauthorized compliance arise only in realistic multi-party environments # multi-agent deployment exposes emergent security vulnerabilities invisible to single-agent evaluation because cross-agent propagation identity spoofing and unauthorized compliance arise only in realistic multi-party environments

View file

@ -1,10 +1,15 @@
--- ---
description: Ben Thompson's structural argument that governments must control frontier AI because it constitutes weapons-grade capability, as demonstrated by the Pentagon's actions against Anthropic description: Ben Thompson's structural argument that governments must control frontier AI because it constitutes weapons-grade capability, as demonstrated by the Pentagon's actions against Anthropic
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-03-06 created: 2026-03-06
source: "Noah Smith, 'If AI is a weapon, why don't we regulate it like one?' (Noahopinion, Mar 6, 2026); Ben Thompson, Stratechery analysis of Anthropic/Pentagon dispute (2026)" source: "Noah Smith, 'If AI is a weapon, why don't we regulate it like one?' (Noahopinion, Mar 6, 2026); Ben Thompson, Stratechery analysis of Anthropic/Pentagon dispute (2026)"
confidence: experimental confidence: experimental
supports:
- "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for"
reweave_edges:
- "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for|supports|2026-03-28"
--- ---
# nation-states will inevitably assert control over frontier AI development because the monopoly on force is the foundational state function and weapons-grade AI capability in private hands is structurally intolerable to governments # nation-states will inevitably assert control over frontier AI development because the monopoly on force is the foundational state function and weapons-grade AI capability in private hands is structurally intolerable to governments

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "UK research strategy identifies human agency, security, privacy, transparency, fairness, value alignment, and accountability as necessary trust conditions" description: "UK research strategy identifies human agency, security, privacy, transparency, fairness, value alignment, and accountability as necessary trust conditions"
@ -6,6 +7,10 @@ confidence: experimental
source: "UK AI for CI Research Network, Artificial Intelligence for Collective Intelligence: A National-Scale Research Strategy (2024)" source: "UK AI for CI Research Network, Artificial Intelligence for Collective Intelligence: A National-Scale Research Strategy (2024)"
created: 2026-03-11 created: 2026-03-11
secondary_domains: [collective-intelligence, critical-systems] secondary_domains: [collective-intelligence, critical-systems]
related:
- "ai enhanced collective intelligence requires federated learning architectures to preserve data sovereignty at scale"
reweave_edges:
- "ai enhanced collective intelligence requires federated learning architectures to preserve data sovereignty at scale|related|2026-03-28"
--- ---
# National-scale collective intelligence infrastructure requires seven trust properties to achieve legitimacy # National-scale collective intelligence infrastructure requires seven trust properties to achieve legitimacy

View file

@ -1,10 +1,21 @@
--- ---
description: Current alignment approaches are all single-model focused while the hardest problems preference diversity scalable oversight and value evolution are inherently collective description: Current alignment approaches are all single-model focused while the hardest problems preference diversity scalable oversight and value evolution are inherently collective
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-17 created: 2026-02-17
source: "Survey of alignment research landscape 2025-2026" source: "Survey of alignment research landscape 2025-2026"
confidence: likely confidence: likely
related:
- "ai enhanced collective intelligence requires federated learning architectures to preserve data sovereignty at scale"
- "national scale collective intelligence infrastructure requires seven trust properties to achieve legitimacy"
- "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach"
reweave_edges:
- "ai enhanced collective intelligence requires federated learning architectures to preserve data sovereignty at scale|related|2026-03-28"
- "national scale collective intelligence infrastructure requires seven trust properties to achieve legitimacy|related|2026-03-28"
- "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach|related|2026-03-28"
--- ---
# no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it # no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Comprehensive review of AI governance mechanisms (2023-2026) shows only the EU AI Act, China's AI regulations, and US export controls produced verified behavioral change at frontier labs — all voluntary mechanisms failed" description: "Comprehensive review of AI governance mechanisms (2023-2026) shows only the EU AI Act, China's AI regulations, and US export controls produced verified behavioral change at frontier labs — all voluntary mechanisms failed"
confidence: likely confidence: likely
source: "Stanford FMTI (Dec 2025), EU enforcement actions (2025), TIME/CNN on Anthropic RSP (Feb 2026), TechCrunch on OpenAI Preparedness Framework (Apr 2025), Fortune on Seoul violations (Aug 2025), Brookings analysis, OECD reports; theseus AI coordination research (Mar 2026)" source: "Stanford FMTI (Dec 2025), EU enforcement actions (2025), TIME/CNN on Anthropic RSP (Feb 2026), TechCrunch on OpenAI Preparedness Framework (Apr 2025), Fortune on Seoul violations (Aug 2025), Brookings analysis, OECD reports; theseus AI coordination research (Mar 2026)"
created: 2026-03-16 created: 2026-03-16
related:
- "UK AI Safety Institute"
reweave_edges:
- "UK AI Safety Institute|related|2026-03-28"
--- ---
# only binding regulation with enforcement teeth changes frontier AI lab behavior because every voluntary commitment has been eroded abandoned or made conditional on competitor behavior when commercially inconvenient # only binding regulation with enforcement teeth changes frontier AI lab behavior because every voluntary commitment has been eroded abandoned or made conditional on competitor behavior when commercially inconvenient

View file

@ -1,10 +1,15 @@
--- ---
description: Some disagreements cannot be resolved with more evidence because they stem from genuine value differences or incommensurable goods and systems must map rather than eliminate them description: Some disagreements cannot be resolved with more evidence because they stem from genuine value differences or incommensurable goods and systems must map rather than eliminate them
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-03-02 created: 2026-03-02
confidence: likely confidence: likely
source: "Arrow's impossibility theorem; value pluralism (Isaiah Berlin); LivingIP design principles" source: "Arrow's impossibility theorem; value pluralism (Isaiah Berlin); LivingIP design principles"
supports:
- "pluralistic ai alignment through multiple systems preserves value diversity better than forced consensus"
reweave_edges:
- "pluralistic ai alignment through multiple systems preserves value diversity better than forced consensus|supports|2026-03-28"
--- ---
# persistent irreducible disagreement # persistent irreducible disagreement

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "CoWoS packaging, HBM memory, and datacenter power each gate AI compute scaling on timescales (2-10 years) much longer than algorithmic or architectural advances (months) — this mismatch creates a window where alignment research can outpace deployment even without deliberate slowdown" description: "CoWoS packaging, HBM memory, and datacenter power each gate AI compute scaling on timescales (2-10 years) much longer than algorithmic or architectural advances (months) — this mismatch creates a window where alignment research can outpace deployment even without deliberate slowdown"
@ -14,6 +15,10 @@ challenged_by:
- "If the US self-limits via infrastructure lag, compute migrates to jurisdictions with fewer safety norms" - "If the US self-limits via infrastructure lag, compute migrates to jurisdictions with fewer safety norms"
secondary_domains: secondary_domains:
- collective-intelligence - collective-intelligence
related:
- "inference efficiency gains erode AI deployment governance without triggering compute monitoring thresholds because governance frameworks target training concentration while inference optimization distributes capability below detection"
reweave_edges:
- "inference efficiency gains erode AI deployment governance without triggering compute monitoring thresholds because governance frameworks target training concentration while inference optimization distributes capability below detection|related|2026-03-28"
--- ---
# Physical infrastructure constraints on AI scaling create a natural governance window because packaging memory and power bottlenecks operate on 2-10 year timescales while capability research advances in months # Physical infrastructure constraints on AI scaling create a natural governance window because packaging memory and power bottlenecks operate on 2-10 year timescales while capability research advances in months

View file

@ -1,10 +1,25 @@
--- ---
description: Three forms of alignment pluralism -- Overton steerable and distributional -- are needed because standard alignment procedures actively reduce the diversity of model outputs description: Three forms of alignment pluralism -- Overton steerable and distributional -- are needed because standard alignment procedures actively reduce the diversity of model outputs
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-17 created: 2026-02-17
source: "Sorensen et al, Roadmap to Pluralistic Alignment (arXiv 2402.05070, ICML 2024); Klassen et al, Pluralistic Alignment Over Time (arXiv 2411.10654, NeurIPS 2024); Harland et al, Adaptive Alignment (arXiv 2410.23630, NeurIPS 2024)" source: "Sorensen et al, Roadmap to Pluralistic Alignment (arXiv 2402.05070, ICML 2024); Klassen et al, Pluralistic Alignment Over Time (arXiv 2411.10654, NeurIPS 2024); Harland et al, Adaptive Alignment (arXiv 2410.23630, NeurIPS 2024)"
confidence: likely confidence: likely
related:
- "minority preference alignment improves 33 percent without majority compromise suggesting single reward leaves value on table"
- "the variance of a learned preference sensitivity distribution diagnoses dataset heterogeneity and collapses to fixed parameter behavior when preferences are homogeneous"
reweave_edges:
- "minority preference alignment improves 33 percent without majority compromise suggesting single reward leaves value on table|related|2026-03-28"
- "pluralistic ai alignment through multiple systems preserves value diversity better than forced consensus|supports|2026-03-28"
- "single reward rlhf cannot align diverse preferences because alignment gap grows proportional to minority distinctiveness|supports|2026-03-28"
- "the variance of a learned preference sensitivity distribution diagnoses dataset heterogeneity and collapses to fixed parameter behavior when preferences are homogeneous|related|2026-03-28"
supports:
- "pluralistic ai alignment through multiple systems preserves value diversity better than forced consensus"
- "single reward rlhf cannot align diverse preferences because alignment gap grows proportional to minority distinctiveness"
--- ---
# pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state # pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state

View file

@ -1,10 +1,19 @@
--- ---
description: The intelligence explosion dynamic occurs when an AI crosses the threshold where it can improve itself faster than humans can, creating a self-reinforcing feedback loop description: The intelligence explosion dynamic occurs when an AI crosses the threshold where it can improve itself faster than humans can, creating a self-reinforcing feedback loop
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-02-16 created: 2026-02-16
source: "Bostrom, Superintelligence: Paths, Dangers, Strategies (2014)" source: "Bostrom, Superintelligence: Paths, Dangers, Strategies (2014)"
confidence: likely confidence: likely
supports:
- "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation"
reweave_edges:
- "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation|supports|2026-03-28"
- "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power|related|2026-03-28"
related:
- "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power"
--- ---
Bostrom formalizes the dynamics of an intelligence explosion using two variables: optimization power (quality-weighted design effort applied to increase the system's intelligence) and recalcitrance (the inverse of the system's responsiveness to that effort). The rate of change in intelligence equals optimization power divided by recalcitrance. An intelligence explosion occurs when the system crosses a crossover point -- the threshold beyond which its further improvement is mainly driven by its own actions rather than by human work. Bostrom formalizes the dynamics of an intelligence explosion using two variables: optimization power (quality-weighted design effort applied to increase the system's intelligence) and recalcitrance (the inverse of the system's responsiveness to that effort). The rate of change in intelligence equals optimization power divided by recalcitrance. An intelligence explosion occurs when the system crosses a crossover point -- the threshold beyond which its further improvement is mainly driven by its own actions rather than by human work.

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [mechanisms] secondary_domains: [mechanisms]
@ -6,6 +8,13 @@ description: "The aggregated rankings variant of RLCHF applies formal social cho
confidence: experimental confidence: experimental
source: "Conitzer et al. (2024), 'Social Choice Should Guide AI Alignment' (ICML 2024)" source: "Conitzer et al. (2024), 'Social Choice Should Guide AI Alignment' (ICML 2024)"
created: 2026-03-11 created: 2026-03-11
related:
- "rlchf features based variant models individual preferences with evaluator characteristics enabling aggregation across diverse groups"
reweave_edges:
- "rlchf features based variant models individual preferences with evaluator characteristics enabling aggregation across diverse groups|related|2026-03-28"
- "rlhf is implicit social choice without normative scrutiny|supports|2026-03-28"
supports:
- "rlhf is implicit social choice without normative scrutiny"
--- ---
# RLCHF aggregated rankings variant combines evaluator rankings via social welfare function before reward model training # RLCHF aggregated rankings variant combines evaluator rankings via social welfare function before reward model training

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [mechanisms] secondary_domains: [mechanisms]
@ -6,6 +7,10 @@ description: "The features-based RLCHF variant learns individual preference mode
confidence: experimental confidence: experimental
source: "Conitzer et al. (2024), 'Social Choice Should Guide AI Alignment' (ICML 2024)" source: "Conitzer et al. (2024), 'Social Choice Should Guide AI Alignment' (ICML 2024)"
created: 2026-03-11 created: 2026-03-11
related:
- "rlchf aggregated rankings variant combines evaluator rankings via social welfare function before reward model training"
reweave_edges:
- "rlchf aggregated rankings variant combines evaluator rankings via social welfare function before reward model training|related|2026-03-28"
--- ---
# RLCHF features-based variant models individual preferences with evaluator characteristics enabling aggregation across diverse groups # RLCHF features-based variant models individual preferences with evaluator characteristics enabling aggregation across diverse groups

View file

@ -1,10 +1,25 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Current RLHF implementations make social choice decisions about evaluator selection and preference aggregation without examining their normative properties" description: "Current RLHF implementations make social choice decisions about evaluator selection and preference aggregation without examining their normative properties"
confidence: likely confidence: likely
source: "Conitzer et al. (2024), 'Social Choice Should Guide AI Alignment' (ICML 2024)" source: "Conitzer et al. (2024), 'Social Choice Should Guide AI Alignment' (ICML 2024)"
created: 2026-03-11 created: 2026-03-11
related:
- "maxmin rlhf applies egalitarian social choice to alignment by maximizing minimum utility across preference groups"
- "rlchf aggregated rankings variant combines evaluator rankings via social welfare function before reward model training"
- "rlchf features based variant models individual preferences with evaluator characteristics enabling aggregation across diverse groups"
reweave_edges:
- "maxmin rlhf applies egalitarian social choice to alignment by maximizing minimum utility across preference groups|related|2026-03-28"
- "representative sampling and deliberative mechanisms should replace convenience platforms for ai alignment feedback|supports|2026-03-28"
- "rlchf aggregated rankings variant combines evaluator rankings via social welfare function before reward model training|related|2026-03-28"
- "rlchf features based variant models individual preferences with evaluator characteristics enabling aggregation across diverse groups|related|2026-03-28"
supports:
- "representative sampling and deliberative mechanisms should replace convenience platforms for ai alignment feedback"
--- ---
# RLHF is implicit social choice without normative scrutiny # RLHF is implicit social choice without normative scrutiny

View file

@ -1,10 +1,25 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Formal impossibility result showing single reward models fail when human preferences are diverse across subpopulations" description: "Formal impossibility result showing single reward models fail when human preferences are diverse across subpopulations"
confidence: likely confidence: likely
source: "Chakraborty et al., MaxMin-RLHF: Alignment with Diverse Human Preferences (ICML 2024)" source: "Chakraborty et al., MaxMin-RLHF: Alignment with Diverse Human Preferences (ICML 2024)"
created: 2026-03-11 created: 2026-03-11
supports:
- "maxmin rlhf applies egalitarian social choice to alignment by maximizing minimum utility across preference groups"
- "minority preference alignment improves 33 percent without majority compromise suggesting single reward leaves value on table"
- "rlchf features based variant models individual preferences with evaluator characteristics enabling aggregation across diverse groups"
reweave_edges:
- "maxmin rlhf applies egalitarian social choice to alignment by maximizing minimum utility across preference groups|supports|2026-03-28"
- "minority preference alignment improves 33 percent without majority compromise suggesting single reward leaves value on table|supports|2026-03-28"
- "rlchf features based variant models individual preferences with evaluator characteristics enabling aggregation across diverse groups|supports|2026-03-28"
- "rlhf is implicit social choice without normative scrutiny|related|2026-03-28"
related:
- "rlhf is implicit social choice without normative scrutiny"
--- ---
# Single-reward RLHF cannot align diverse preferences because alignment gap grows proportional to minority distinctiveness and inversely to representation # Single-reward RLHF cannot align diverse preferences because alignment gap grows proportional to minority distinctiveness and inversely to representation

View file

@ -1,10 +1,15 @@
--- ---
description: Some disagreements cannot be resolved with more evidence because they stem from genuine value differences or incommensurable goods and systems must map rather than eliminate them description: Some disagreements cannot be resolved with more evidence because they stem from genuine value differences or incommensurable goods and systems must map rather than eliminate them
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-03-02 created: 2026-03-02
confidence: likely confidence: likely
source: "Arrow's impossibility theorem; value pluralism (Isaiah Berlin); LivingIP design principles" source: "Arrow's impossibility theorem; value pluralism (Isaiah Berlin); LivingIP design principles"
supports:
- "pluralistic ai alignment through multiple systems preserves value diversity better than forced consensus"
reweave_edges:
- "pluralistic ai alignment through multiple systems preserves value diversity better than forced consensus|supports|2026-03-28"
--- ---
# some disagreements are permanently irreducible because they stem from genuine value differences not information gaps and systems must map rather than eliminate them # some disagreements are permanently irreducible because they stem from genuine value differences not information gaps and systems must map rather than eliminate them

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "AI coding tools evolve through distinct stages (autocomplete → single agent → parallel agents → agent teams) and each stage has an optimal adoption frontier where moving too aggressively nets chaos while moving too conservatively wastes leverage" description: "AI coding tools evolve through distinct stages (autocomplete → single agent → parallel agents → agent teams) and each stage has an optimal adoption frontier where moving too aggressively nets chaos while moving too conservatively wastes leverage"
confidence: likely confidence: likely
source: "Andrej Karpathy (@karpathy), analysis of Cursor tab-to-agent ratio data, Feb 2026" source: "Andrej Karpathy (@karpathy), analysis of Cursor tab-to-agent ratio data, Feb 2026"
created: 2026-03-09 created: 2026-03-09
related:
- "as AI automated software development becomes certain the bottleneck shifts from building capacity to knowing what to build making structured knowledge graphs the critical input to autonomous systems"
reweave_edges:
- "as AI automated software development becomes certain the bottleneck shifts from building capacity to knowing what to build making structured knowledge graphs the critical input to autonomous systems|related|2026-03-28"
--- ---
# The progression from autocomplete to autonomous agent teams follows a capability-matched escalation where premature adoption creates more chaos than value # The progression from autocomplete to autonomous agent teams follows a capability-matched escalation where premature adoption creates more chaos than value

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence] secondary_domains: [collective-intelligence]
@ -6,6 +8,13 @@ description: "The Residue prompt applied identically to GPT-5.4 Thinking and Cla
confidence: experimental confidence: experimental
source: "Aquino-Michaels 2026, 'Completing Claude's Cycles' (github.com/no-way-labs/residue), meta_log.md and agent logs" source: "Aquino-Michaels 2026, 'Completing Claude's Cycles' (github.com/no-way-labs/residue), meta_log.md and agent logs"
created: 2026-03-07 created: 2026-03-07
related:
- "AI agents excel at implementing well scoped ideas but cannot generate creative experiment designs which makes the human role shift from researcher to agent workflow architect"
reweave_edges:
- "AI agents excel at implementing well scoped ideas but cannot generate creative experiment designs which makes the human role shift from researcher to agent workflow architect|related|2026-03-28"
- "tools and artifacts transfer between AI agents and evolve in the process because Agent O improved Agent Cs solver by combining it with its own structural knowledge creating a hybrid better than either original|supports|2026-03-28"
supports:
- "tools and artifacts transfer between AI agents and evolve in the process because Agent O improved Agent Cs solver by combining it with its own structural knowledge creating a hybrid better than either original"
--- ---
# the same coordination protocol applied to different AI models produces radically different problem-solving strategies because the protocol structures process not thought # the same coordination protocol applied to different AI models produces radically different problem-solving strategies because the protocol structures process not thought

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "As inference grows from ~33% to ~66% of AI compute by 2026, the hardware landscape shifts from NVIDIA-monopolized centralized training clusters to diverse distributed inference on ARM, custom ASICs, and edge devices — changing who can deploy AI capability and how governable deployment is" description: "As inference grows from ~33% to ~66% of AI compute by 2026, the hardware landscape shifts from NVIDIA-monopolized centralized training clusters to diverse distributed inference on ARM, custom ASICs, and edge devices — changing who can deploy AI capability and how governable deployment is"
@ -14,6 +15,10 @@ challenged_by:
- "Inference at scale (serving billions of users) still requires massive centralized infrastructure" - "Inference at scale (serving billions of users) still requires massive centralized infrastructure"
secondary_domains: secondary_domains:
- collective-intelligence - collective-intelligence
supports:
- "inference efficiency gains erode AI deployment governance without triggering compute monitoring thresholds because governance frameworks target training concentration while inference optimization distributes capability below detection"
reweave_edges:
- "inference efficiency gains erode AI deployment governance without triggering compute monitoring thresholds because governance frameworks target training concentration while inference optimization distributes capability below detection|supports|2026-03-28"
--- ---
# The training-to-inference shift structurally favors distributed AI architectures because inference optimizes for power efficiency and cost-per-token where diverse hardware competes while training optimizes for raw throughput where NVIDIA monopolizes # The training-to-inference shift structurally favors distributed AI architectures because inference optimizes for power efficiency and cost-per-token where diverse hardware competes while training optimizes for raw throughput where NVIDIA monopolizes

View file

@ -1,10 +1,15 @@
--- ---
description: Noah Smith argues that cognitive superintelligence alone cannot produce AI takeover — physical autonomy, robotics, and full production chain control are necessary preconditions, none of which current AI possesses description: Noah Smith argues that cognitive superintelligence alone cannot produce AI takeover — physical autonomy, robotics, and full production chain control are necessary preconditions, none of which current AI possesses
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-03-06 created: 2026-03-06
source: "Noah Smith, 'Superintelligence is already here, today' (Noahopinion, Mar 2, 2026)" source: "Noah Smith, 'Superintelligence is already here, today' (Noahopinion, Mar 2, 2026)"
confidence: experimental confidence: experimental
related:
- "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power"
reweave_edges:
- "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power|related|2026-03-28"
--- ---
# three conditions gate AI takeover risk autonomy robotics and production chain control and current AI satisfies none of them which bounds near-term catastrophic risk despite superhuman cognitive capabilities # three conditions gate AI takeover risk autonomy robotics and production chain control and current AI satisfies none of them which bounds near-term catastrophic risk despite superhuman cognitive capabilities

View file

@ -1,10 +1,15 @@
--- ---
description: Anthropic's Feb 2026 rollback of its Responsible Scaling Policy proves that even the strongest voluntary safety commitment collapses when the competitive cost exceeds the reputational benefit description: Anthropic's Feb 2026 rollback of its Responsible Scaling Policy proves that even the strongest voluntary safety commitment collapses when the competitive cost exceeds the reputational benefit
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-03-06 created: 2026-03-06
source: "Anthropic RSP v3.0 (Feb 24, 2026); TIME exclusive (Feb 25, 2026); Jared Kaplan statements" source: "Anthropic RSP v3.0 (Feb 24, 2026); TIME exclusive (Feb 25, 2026); Jared Kaplan statements"
confidence: likely confidence: likely
supports:
- "Anthropic"
reweave_edges:
- "Anthropic|supports|2026-03-28"
--- ---
# voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints # voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: collective-intelligence domain: collective-intelligence
description: "When agents share aspects of their generative models they can pursue collective goals without negotiating individual contributions" description: "When agents share aspects of their generative models they can pursue collective goals without negotiating individual contributions"
@ -7,6 +8,10 @@ source: "Albarracin et al., 'Shared Protentions in Multi-Agent Active Inference'
created: 2026-03-11 created: 2026-03-11
secondary_domains: [ai-alignment] secondary_domains: [ai-alignment]
depends_on: ["shared-anticipatory-structures-enable-decentralized-coordination"] depends_on: ["shared-anticipatory-structures-enable-decentralized-coordination"]
supports:
- "factorised generative models enable decentralized multi agent representation through individual level beliefs"
reweave_edges:
- "factorised generative models enable decentralized multi agent representation through individual level beliefs|supports|2026-03-28"
--- ---
# Shared generative models enable implicit coordination through shared predictions rather than explicit communication or hierarchy # Shared generative models enable implicit coordination through shared predictions rather than explicit communication or hierarchy

View file

@ -1,10 +1,15 @@
--- ---
description: 173 AI-discovered programs now in clinical development with 80-90 percent Phase I success and Insilicos rentosertib is first fully AI-designed drug to clear Phase IIa but overall clinical failure rates remain unchanged making later-stage success the key unknown description: 173 AI-discovered programs now in clinical development with 80-90 percent Phase I success and Insilicos rentosertib is first fully AI-designed drug to clear Phase IIa but overall clinical failure rates remain unchanged making later-stage success the key unknown
type: claim type: claim
domain: health domain: health
created: 2026-02-17 created: 2026-02-17
source: "AI drug discovery pipeline data 2026; Insilico Medicine rentosertib Phase IIa; Isomorphic Labs $3B partnerships; WEF drug discovery analysis January 2026" source: "AI drug discovery pipeline data 2026; Insilico Medicine rentosertib Phase IIa; Isomorphic Labs $3B partnerships; WEF drug discovery analysis January 2026"
confidence: likely confidence: likely
related:
- "FDA is replacing animal testing with AI models and organ on chip as the default preclinical pathway which will compress drug development timelines and reduce the 90 percent clinical failure rate"
reweave_edges:
- "FDA is replacing animal testing with AI models and organ on chip as the default preclinical pathway which will compress drug development timelines and reduce the 90 percent clinical failure rate|related|2026-03-28"
--- ---
# AI compresses drug discovery timelines by 30-40 percent but has not yet improved the 90 percent clinical failure rate that determines industry economics # AI compresses drug discovery timelines by 30-40 percent but has not yet improved the 90 percent clinical failure rate that determines industry economics

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: health domain: health
description: "92% of US health systems deploying AI scribes by March 2025 — a 2-3 year adoption curve vs 15 years for EHRs — because documentation is the one clinical workflow where AI improvement is immediately measurable, carries minimal patient risk, and delivers revenue capture gains" description: "92% of US health systems deploying AI scribes by March 2025 — a 2-3 year adoption curve vs 15 years for EHRs — because documentation is the one clinical workflow where AI improvement is immediately measurable, carries minimal patient risk, and delivers revenue capture gains"
confidence: proven confidence: proven
source: "Bessemer Venture Partners, State of Health AI 2026 (bvp.com/atlas/state-of-health-ai-2026)" source: "Bessemer Venture Partners, State of Health AI 2026 (bvp.com/atlas/state-of-health-ai-2026)"
created: 2026-03-07 created: 2026-03-07
related:
- "AI native health companies achieve 3 5x the revenue productivity of traditional health services because AI eliminates the linear scaling constraint between headcount and output"
reweave_edges:
- "AI native health companies achieve 3 5x the revenue productivity of traditional health services because AI eliminates the linear scaling constraint between headcount and output|related|2026-03-28"
--- ---
# AI scribes reached 92 percent provider adoption in under 3 years because documentation is the rare healthcare workflow where AI value is immediate unambiguous and low-risk # AI scribes reached 92 percent provider adoption in under 3 years because documentation is the rare healthcare workflow where AI value is immediate unambiguous and low-risk

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: health domain: health
description: "CMS adding category I CPT codes for AI-assisted diagnosis (diabetic retinopathy, coronary plaque) and testing category III codes for AI ECG, echocardiograms, and ultrasound — creating the first formal reimbursement pathway for clinical AI" description: "CMS adding category I CPT codes for AI-assisted diagnosis (diabetic retinopathy, coronary plaque) and testing category III codes for AI ECG, echocardiograms, and ultrasound — creating the first formal reimbursement pathway for clinical AI"
confidence: likely confidence: likely
source: "Bessemer Venture Partners, State of Health AI 2026 (bvp.com/atlas/state-of-health-ai-2026)" source: "Bessemer Venture Partners, State of Health AI 2026 (bvp.com/atlas/state-of-health-ai-2026)"
created: 2026-03-07 created: 2026-03-07
supports:
- "consumer willingness to pay out of pocket for AI enhanced care is outpacing reimbursement creating a cash pay adoption pathway that bypasses traditional payer gatekeeping"
reweave_edges:
- "consumer willingness to pay out of pocket for AI enhanced care is outpacing reimbursement creating a cash pay adoption pathway that bypasses traditional payer gatekeeping|supports|2026-03-28"
--- ---
# CMS is creating AI-specific reimbursement codes which will formalize a two-speed adoption system where proven AI applications get payment parity while experimental ones remain in cash-pay limbo # CMS is creating AI-specific reimbursement codes which will formalize a two-speed adoption system where proven AI applications get payment parity while experimental ones remain in cash-pay limbo

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: health domain: health
description: "Universal workforce shortages and facility closures indicate systemic care capacity failure not regional variation" description: "Universal workforce shortages and facility closures indicate systemic care capacity failure not regional variation"
confidence: proven confidence: proven
source: "AARP 2025 Caregiving Report" source: "AARP 2025 Caregiving Report"
created: 2026-03-11 created: 2026-03-11
supports:
- "family caregiving functions as poverty transmission mechanism forcing debt savings depletion and food insecurity on working age population"
reweave_edges:
- "family caregiving functions as poverty transmission mechanism forcing debt savings depletion and food insecurity on working age population|supports|2026-03-28"
--- ---
# Caregiver workforce crisis shows all 50 states experiencing shortages with 43 states reporting facility closures signaling care infrastructure collapse # Caregiver workforce crisis shows all 50 states experiencing shortages with 43 states reporting facility closures signaling care infrastructure collapse

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: health domain: health
description: "RadNet's AI mammography study shows 36% of women paying $40 out-of-pocket for AI screening with 43% higher cancer detection, suggesting consumer demand will drive AI adoption faster than CMS reimbursement codes" description: "RadNet's AI mammography study shows 36% of women paying $40 out-of-pocket for AI screening with 43% higher cancer detection, suggesting consumer demand will drive AI adoption faster than CMS reimbursement codes"
confidence: likely confidence: likely
source: "Bessemer Venture Partners, State of Health AI 2026 (bvp.com/atlas/state-of-health-ai-2026)" source: "Bessemer Venture Partners, State of Health AI 2026 (bvp.com/atlas/state-of-health-ai-2026)"
created: 2026-03-07 created: 2026-03-07
related:
- "CMS is creating AI specific reimbursement codes which will formalize a two speed adoption system where proven AI applications get payment parity while experimental ones remain in cash pay limbo"
reweave_edges:
- "CMS is creating AI specific reimbursement codes which will formalize a two speed adoption system where proven AI applications get payment parity while experimental ones remain in cash pay limbo|related|2026-03-28"
--- ---
# consumer willingness to pay out of pocket for AI-enhanced care is outpacing reimbursement creating a cash-pay adoption pathway that bypasses traditional payer gatekeeping # consumer willingness to pay out of pocket for AI-enhanced care is outpacing reimbursement creating a cash-pay adoption pathway that bypasses traditional payer gatekeeping

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: health domain: health
description: "Unpaid care responsibilities transfer elderly health costs to working-age families through financial sacrifice that compounds over decades" description: "Unpaid care responsibilities transfer elderly health costs to working-age families through financial sacrifice that compounds over decades"
confidence: likely confidence: likely
source: "AARP 2025 Caregiving Report" source: "AARP 2025 Caregiving Report"
created: 2026-03-11 created: 2026-03-11
supports:
- "caregiver workforce crisis shows all 50 states experiencing shortages with 43 states reporting facility closures signaling care infrastructure collapse"
reweave_edges:
- "caregiver workforce crisis shows all 50 states experiencing shortages with 43 states reporting facility closures signaling care infrastructure collapse|supports|2026-03-28"
--- ---
# Family caregiving functions as poverty transmission mechanism forcing debt savings depletion and food insecurity on working-age population # Family caregiving functions as poverty transmission mechanism forcing debt savings depletion and food insecurity on working-age population

View file

@ -1,10 +1,15 @@
--- ---
description: Current gene therapies cost 2-4 million dollars per treatment using ex vivo editing but in vivo approaches like Verve's one-time PCSK9 base editing infusion showing 53 percent LDL reduction could reach 50-200K by 2035 making curative medicine scalable description: Current gene therapies cost 2-4 million dollars per treatment using ex vivo editing but in vivo approaches like Verve's one-time PCSK9 base editing infusion showing 53 percent LDL reduction could reach 50-200K by 2035 making curative medicine scalable
type: claim type: claim
domain: health domain: health
created: 2026-02-17 created: 2026-02-17
source: "IGI CRISPR clinical trials update 2025; BioPharma Dive Verve PCSK9 data; BioInformant FDA-approved CGT database; GEN reimbursement outlook 2025; PMC gene therapy pipeline analysis" source: "IGI CRISPR clinical trials update 2025; BioPharma Dive Verve PCSK9 data; BioInformant FDA-approved CGT database; GEN reimbursement outlook 2025; PMC gene therapy pipeline analysis"
confidence: likely confidence: likely
related:
- "FDA is replacing animal testing with AI models and organ on chip as the default preclinical pathway which will compress drug development timelines and reduce the 90 percent clinical failure rate"
reweave_edges:
- "FDA is replacing animal testing with AI models and organ on chip as the default preclinical pathway which will compress drug development timelines and reduce the 90 percent clinical failure rate|related|2026-03-28"
--- ---
# gene editing is shifting from ex vivo to in vivo delivery via lipid nanoparticles which will reduce curative therapy costs from millions to hundreds of thousands per treatment # gene editing is shifting from ex vivo to in vivo delivery via lipid nanoparticles which will reduce curative therapy costs from millions to hundreds of thousands per treatment

View file

@ -1,10 +1,21 @@
--- ---
description: Nearly every AI application in healthcare optimizes the 10-20% clinical side while 80-90% of outcomes are driven by non-clinical factors so making sick care more efficient produces more sick care not better health description: Nearly every AI application in healthcare optimizes the 10-20% clinical side while 80-90% of outcomes are driven by non-clinical factors so making sick care more efficient produces more sick care not better health
type: claim type: claim
domain: health domain: health
created: 2026-02-23 created: 2026-02-23
source: "Devoted Health AI Overview Memo, 2026" source: "Devoted Health AI Overview Memo, 2026"
confidence: likely confidence: likely
related:
- "AI native health companies achieve 3 5x the revenue productivity of traditional health services because AI eliminates the linear scaling constraint between headcount and output"
- "CMS is creating AI specific reimbursement codes which will formalize a two speed adoption system where proven AI applications get payment parity while experimental ones remain in cash pay limbo"
- "consumer willingness to pay out of pocket for AI enhanced care is outpacing reimbursement creating a cash pay adoption pathway that bypasses traditional payer gatekeeping"
reweave_edges:
- "AI native health companies achieve 3 5x the revenue productivity of traditional health services because AI eliminates the linear scaling constraint between headcount and output|related|2026-03-28"
- "CMS is creating AI specific reimbursement codes which will formalize a two speed adoption system where proven AI applications get payment parity while experimental ones remain in cash pay limbo|related|2026-03-28"
- "consumer willingness to pay out of pocket for AI enhanced care is outpacing reimbursement creating a cash pay adoption pathway that bypasses traditional payer gatekeeping|related|2026-03-28"
--- ---
# healthcare AI creates a Jevons paradox because adding capacity to sick care induces more demand for sick care # healthcare AI creates a Jevons paradox because adding capacity to sick care induces more demand for sick care

View file

@ -1,10 +1,15 @@
--- ---
description: Global healthcare venture financing reached 60.4 billion in 2025 but AI-native companies capture 54 percent of funding with a 19 percent deal premium while mega-deals over 100 million account for 42 percent of total and Agilon collapsed from 10 billion to 255 million description: Global healthcare venture financing reached 60.4 billion in 2025 but AI-native companies capture 54 percent of funding with a 19 percent deal premium while mega-deals over 100 million account for 42 percent of total and Agilon collapsed from 10 billion to 255 million
type: claim type: claim
domain: health domain: health
created: 2026-02-17 created: 2026-02-17
source: "Health tech VC landscape analysis February 2026; OpenEvidence Abridge Hippocratic AI fundraising disclosures; Agilon Health SEC filings; Rock Health digital health funding reports 2025; Bessemer Venture Partners State of Health AI 2026" source: "Health tech VC landscape analysis February 2026; OpenEvidence Abridge Hippocratic AI fundraising disclosures; Agilon Health SEC filings; Rock Health digital health funding reports 2025; Bessemer Venture Partners State of Health AI 2026"
confidence: likely confidence: likely
related:
- "AI native health companies achieve 3 5x the revenue productivity of traditional health services because AI eliminates the linear scaling constraint between headcount and output"
reweave_edges:
- "AI native health companies achieve 3 5x the revenue productivity of traditional health services because AI eliminates the linear scaling constraint between headcount and output|related|2026-03-28"
--- ---
# healthcare AI funding follows a winner-take-most pattern with category leaders absorbing capital at unprecedented velocity while 35 percent of deals are flat or down rounds # healthcare AI funding follows a winner-take-most pattern with category leaders absorbing capital at unprecedented velocity while 35 percent of deals are flat or down rounds

View file

@ -1,10 +1,21 @@
--- ---
description: Wachter argues AI should be regulated more like physician licensing with competency exams and ongoing certification rather than the FDA approval model designed for drugs and devices that remain static forever description: Wachter argues AI should be regulated more like physician licensing with competency exams and ongoing certification rather than the FDA approval model designed for drugs and devices that remain static forever
type: claim type: claim
domain: health domain: health
created: 2026-02-18 created: 2026-02-18
source: "DJ Patil interviewing Bob Wachter, Commonwealth Club, February 9 2026; Wachter 'A Giant Leap' (2026)" source: "DJ Patil interviewing Bob Wachter, Commonwealth Club, February 9 2026; Wachter 'A Giant Leap' (2026)"
confidence: likely confidence: likely
related:
- "CMS is creating AI specific reimbursement codes which will formalize a two speed adoption system where proven AI applications get payment parity while experimental ones remain in cash pay limbo"
- "FDA is replacing animal testing with AI models and organ on chip as the default preclinical pathway which will compress drug development timelines and reduce the 90 percent clinical failure rate"
- "consumer willingness to pay out of pocket for AI enhanced care is outpacing reimbursement creating a cash pay adoption pathway that bypasses traditional payer gatekeeping"
reweave_edges:
- "CMS is creating AI specific reimbursement codes which will formalize a two speed adoption system where proven AI applications get payment parity while experimental ones remain in cash pay limbo|related|2026-03-28"
- "FDA is replacing animal testing with AI models and organ on chip as the default preclinical pathway which will compress drug development timelines and reduce the 90 percent clinical failure rate|related|2026-03-28"
- "consumer willingness to pay out of pocket for AI enhanced care is outpacing reimbursement creating a cash pay adoption pathway that bypasses traditional payer gatekeeping|related|2026-03-28"
--- ---
# healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software # healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: health domain: health
description: "MA enrollment reached 51% in 2023 and 54% by 2025, with CBO projecting 64% by 2034, making traditional Medicare the minority program" description: "MA enrollment reached 51% in 2023 and 54% by 2025, with CBO projecting 64% by 2034, making traditional Medicare the minority program"
confidence: proven confidence: proven
source: "Kaiser Family Foundation, Medicare Advantage in 2025: Enrollment Update and Key Trends (2025)" source: "Kaiser Family Foundation, Medicare Advantage in 2025: Enrollment Update and Key Trends (2025)"
created: 2025-07-24 created: 2025-07-24
supports:
- "chronic condition special needs plans grew 71 percent in one year indicating explosive demand for disease management infrastructure"
reweave_edges:
- "chronic condition special needs plans grew 71 percent in one year indicating explosive demand for disease management infrastructure|supports|2026-03-28"
--- ---
# Medicare Advantage crossed majority enrollment in 2023 marking structural transformation from supplement to dominant program # Medicare Advantage crossed majority enrollment in 2023 marking structural transformation from supplement to dominant program

View file

@ -1,10 +1,15 @@
--- ---
description: The market and state broke traditional power structures by offering people individuality but this severed the intimate social bonds that sustained human wellbeing for millennia creating alienation depression and meaning deficits that economic growth cannot address description: The market and state broke traditional power structures by offering people individuality but this severed the intimate social bonds that sustained human wellbeing for millennia creating alienation depression and meaning deficits that economic growth cannot address
type: claim type: claim
domain: health domain: health
source: "Architectural Investing, Ch. Dark Side of Specialization; Harari (Sapiens); Perlmutter (Brainwash)" source: "Architectural Investing, Ch. Dark Side of Specialization; Harari (Sapiens); Perlmutter (Brainwash)"
confidence: likely confidence: likely
created: 2026-02-28 created: 2026-02-28
related:
- "family caregiving functions as poverty transmission mechanism forcing debt savings depletion and food insecurity on working age population"
reweave_edges:
- "family caregiving functions as poverty transmission mechanism forcing debt savings depletion and food insecurity on working age population|related|2026-03-28"
--- ---
# modernization dismantles family and community structures replacing them with market and state relationships that increase individual freedom but erode psychosocial foundations of wellbeing # modernization dismantles family and community structures replacing them with market and state relationships that increase individual freedom but erode psychosocial foundations of wellbeing

View file

@ -1,10 +1,19 @@
--- ---
type: claim type: claim
domain: health domain: health
description: "Unpaid family care represents 16% of total US health spending yet remains invisible to policy models and capacity planning" description: "Unpaid family care represents 16% of total US health spending yet remains invisible to policy models and capacity planning"
confidence: proven confidence: proven
source: "AARP 2025 Caregiving Report" source: "AARP 2025 Caregiving Report"
created: 2026-03-11 created: 2026-03-11
related:
- "caregiver workforce crisis shows all 50 states experiencing shortages with 43 states reporting facility closures signaling care infrastructure collapse"
reweave_edges:
- "caregiver workforce crisis shows all 50 states experiencing shortages with 43 states reporting facility closures signaling care infrastructure collapse|related|2026-03-28"
- "family caregiving functions as poverty transmission mechanism forcing debt savings depletion and food insecurity on working age population|supports|2026-03-28"
supports:
- "family caregiving functions as poverty transmission mechanism forcing debt savings depletion and food insecurity on working age population"
--- ---
# Unpaid family caregiving provides 870 billion annually representing 16 percent of total US health economy invisible to policy models # Unpaid family caregiving provides 870 billion annually representing 16 percent of total US health economy invisible to policy models

View file

@ -25,6 +25,12 @@ competitors: ["OpenAI", "Google DeepMind", "xAI"]
tracked_by: theseus tracked_by: theseus
created: 2026-03-16 created: 2026-03-16
last_updated: 2026-03-16 last_updated: 2026-03-16
related:
- "Dario Amodei"
- "OpenAI"
reweave_edges:
- "Dario Amodei|related|2026-03-28"
- "OpenAI|related|2026-03-28"
--- ---
# Anthropic # Anthropic

View file

@ -21,6 +21,12 @@ competitors: ["OpenAI", "Anthropic", "xAI"]
tracked_by: theseus tracked_by: theseus
created: 2026-03-16 created: 2026-03-16
last_updated: 2026-03-16 last_updated: 2026-03-16
related:
- "OpenAI"
- "xAI"
reweave_edges:
- "OpenAI|related|2026-03-28"
- "xAI|related|2026-03-28"
--- ---
# Google DeepMind # Google DeepMind

View file

@ -22,6 +22,20 @@ competitors: ["Anthropic", "Google DeepMind", "xAI"]
tracked_by: theseus tracked_by: theseus
created: 2026-03-16 created: 2026-03-16
last_updated: 2026-03-16 last_updated: 2026-03-16
related:
- "Anthropic"
- "Dario Amodei"
- "Google DeepMind"
- "Safe Superintelligence Inc."
- "Thinking Machines Lab"
- "xAI"
reweave_edges:
- "Anthropic|related|2026-03-28"
- "Dario Amodei|related|2026-03-28"
- "Google DeepMind|related|2026-03-28"
- "Safe Superintelligence Inc.|related|2026-03-28"
- "Thinking Machines Lab|related|2026-03-28"
- "xAI|related|2026-03-28"
--- ---
# OpenAI # OpenAI

View file

@ -20,6 +20,12 @@ competitors: ["OpenAI", "Anthropic", "Google DeepMind"]
tracked_by: theseus tracked_by: theseus
created: 2026-03-16 created: 2026-03-16
last_updated: 2026-03-16 last_updated: 2026-03-16
related:
- "Google DeepMind"
- "OpenAI"
reweave_edges:
- "Google DeepMind|related|2026-03-28"
- "OpenAI|related|2026-03-28"
--- ---
# xAI # xAI

View file

@ -1,10 +1,25 @@
--- ---
description: The dominant alignment paradigms share a core limitation -- human preferences are diverse distributional and context-dependent not reducible to one reward function description: The dominant alignment paradigms share a core limitation -- human preferences are diverse distributional and context-dependent not reducible to one reward function
type: claim type: claim
domain: collective-intelligence domain: collective-intelligence
created: 2026-02-17 created: 2026-02-17
source: "DPO Survey 2025 (arXiv 2503.11701)" source: "DPO Survey 2025 (arXiv 2503.11701)"
confidence: likely confidence: likely
related:
- "rlchf aggregated rankings variant combines evaluator rankings via social welfare function before reward model training"
- "rlhf is implicit social choice without normative scrutiny"
- "the variance of a learned preference sensitivity distribution diagnoses dataset heterogeneity and collapses to fixed parameter behavior when preferences are homogeneous"
reweave_edges:
- "rlchf aggregated rankings variant combines evaluator rankings via social welfare function before reward model training|related|2026-03-28"
- "rlhf is implicit social choice without normative scrutiny|related|2026-03-28"
- "single reward rlhf cannot align diverse preferences because alignment gap grows proportional to minority distinctiveness|supports|2026-03-28"
- "the variance of a learned preference sensitivity distribution diagnoses dataset heterogeneity and collapses to fixed parameter behavior when preferences are homogeneous|related|2026-03-28"
supports:
- "single reward rlhf cannot align diverse preferences because alignment gap grows proportional to minority distinctiveness"
--- ---
# RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values # RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values

View file

@ -1,10 +1,15 @@
--- ---
description: Ashby's Law of Requisite Variety, Kauffman's adjacent possible, Page's diversity theorem, and Henrich's Tasmanian regression all prove diversity is a physical law of adaptive systems description: Ashby's Law of Requisite Variety, Kauffman's adjacent possible, Page's diversity theorem, and Henrich's Tasmanian regression all prove diversity is a physical law of adaptive systems
type: claim type: claim
domain: collective-intelligence domain: collective-intelligence
created: 2026-02-16 created: 2026-02-16
confidence: proven confidence: proven
source: "TeleoHumanity Manifesto, Chapter 4" source: "TeleoHumanity Manifesto, Chapter 4"
supports:
- "human ideas naturally converge toward similarity over social learning chains making AI a net diversity injector rather than a homogenizer under high exposure conditions"
reweave_edges:
- "human ideas naturally converge toward similarity over social learning chains making AI a net diversity injector rather than a homogenizer under high exposure conditions|supports|2026-03-28"
--- ---
# collective intelligence requires diversity as a structural precondition not a moral preference # collective intelligence requires diversity as a structural precondition not a moral preference

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: collective-intelligence domain: collective-intelligence
description: "The formal basis for oversight problems: when agents have private information or unobservable actions, principals cannot design contracts that fully align incentives, creating irreducible gaps between intended and actual behavior" description: "The formal basis for oversight problems: when agents have private information or unobservable actions, principals cannot design contracts that fully align incentives, creating irreducible gaps between intended and actual behavior"
confidence: proven confidence: proven
source: "Jensen & Meckling (1976); Akerlof, Market for Lemons (1970); Holmström (1979); Arrow (1963)" source: "Jensen & Meckling (1976); Akerlof, Market for Lemons (1970); Holmström (1979); Arrow (1963)"
created: 2026-03-07 created: 2026-03-07
related:
- "AI agents as personal advocates collapse Coasean transaction costs enabling bottom up coordination at societal scale but catastrophic risks remain non negotiable requiring state enforcement as outer boundary"
reweave_edges:
- "AI agents as personal advocates collapse Coasean transaction costs enabling bottom up coordination at societal scale but catastrophic risks remain non negotiable requiring state enforcement as outer boundary|related|2026-03-28"
--- ---
# principal-agent problems arise whenever one party acts on behalf of another with divergent interests and unobservable effort because information asymmetry makes perfect contracts impossible # principal-agent problems arise whenever one party acts on behalf of another with divergent interests and unobservable effort because information asymmetry makes perfect contracts impossible

View file

@ -1,10 +1,18 @@
--- ---
description: Safety post-training reduces general utility through forgetting creating competitive pressures where organizations eschew safety to gain capability advantages description: Safety post-training reduces general utility through forgetting creating competitive pressures where organizations eschew safety to gain capability advantages
type: claim type: claim
domain: collective-intelligence domain: collective-intelligence
created: 2026-02-17 created: 2026-02-17
source: "AI Safety Forum discussions; multiple alignment researchers 2025" source: "AI Safety Forum discussions; multiple alignment researchers 2025"
confidence: likely confidence: likely
related:
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations"
- "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference"
reweave_edges:
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations|related|2026-03-28"
- "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference|related|2026-03-28"
--- ---
# the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it # the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it

View file

@ -1,4 +1,5 @@
--- ---
description: Globalized supply chains lean healthcare infrastructure and overleveraged financial systems all optimize for efficiency during normal times while accumulating hidden tail risk that materializes catastrophically during shocks description: Globalized supply chains lean healthcare infrastructure and overleveraged financial systems all optimize for efficiency during normal times while accumulating hidden tail risk that materializes catastrophically during shocks
type: claim type: claim
domain: critical-systems domain: critical-systems
@ -6,6 +7,10 @@ source: "Architectural Investing, Ch. Introduction; Taleb (Black Swan)"
confidence: proven confidence: proven
tradition: "complexity economics, risk management, Teleological Investing" tradition: "complexity economics, risk management, Teleological Investing"
created: 2026-02-28 created: 2026-02-28
related:
- "delegating critical infrastructure development to AI creates civilizational fragility because humans lose the ability to understand maintain and fix the systems civilization depends on"
reweave_edges:
- "delegating critical infrastructure development to AI creates civilizational fragility because humans lose the ability to understand maintain and fix the systems civilization depends on|related|2026-03-28"
--- ---
# optimization for efficiency without regard for resilience creates systemic fragility because interconnected systems transmit and amplify local failures into cascading breakdowns # optimization for efficiency without regard for resilience creates systemic fragility because interconnected systems transmit and amplify local failures into cascading breakdowns

View file

@ -1,10 +1,15 @@
--- ---
type: claim type: claim
domain: cultural-dynamics domain: cultural-dynamics
description: "Olson's logic of collective action: large groups systematically underprovide public goods because individual incentives favor free-riding, and this problem worsens with group size — small concentrated groups outorganize large diffuse ones" description: "Olson's logic of collective action: large groups systematically underprovide public goods because individual incentives favor free-riding, and this problem worsens with group size — small concentrated groups outorganize large diffuse ones"
confidence: proven confidence: proven
source: "Olson 1965 The Logic of Collective Action; Ostrom 1990 Governing the Commons (boundary condition)" source: "Olson 1965 The Logic of Collective Action; Ostrom 1990 Governing the Commons (boundary condition)"
created: 2026-03-08 created: 2026-03-08
related:
- "AI agents as personal advocates collapse Coasean transaction costs enabling bottom up coordination at societal scale but catastrophic risks remain non negotiable requiring state enforcement as outer boundary"
reweave_edges:
- "AI agents as personal advocates collapse Coasean transaction costs enabling bottom up coordination at societal scale but catastrophic risks remain non negotiable requiring state enforcement as outer boundary|related|2026-03-28"
--- ---
# collective action fails by default because rational individuals free-ride on group efforts when they cannot be excluded from benefits regardless of contribution # collective action fails by default because rational individuals free-ride on group efforts when they cannot be excluded from benefits regardless of contribution