reweave: connect 30 orphan claims via vector similarity

Threshold: 0.7, Haiku classification, 48 files modified.

Pentagon-Agent: Epimetheus <0144398e-4ed3-4fe2-95a3-3d72e1abf887>
This commit is contained in:
Teleo Agents 2026-04-07 00:14:44 +00:00
parent b6739f718d
commit 6143220b66
48 changed files with 237 additions and 0 deletions

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: living-agents domain: living-agents
description: "The Teleo knowledge base uses wiki links as typed edges in a reasoning graph where claims ground beliefs and beliefs ground positions, creating chains that any agent can audit from conclusion back to evidence" description: "The Teleo knowledge base uses wiki links as typed edges in a reasoning graph where claims ground beliefs and beliefs ground positions, creating chains that any agent can audit from conclusion back to evidence"
@ -7,8 +8,10 @@ source: "Teleo collective operational evidence — belief files cite 3+ claims,
created: 2026-03-07 created: 2026-03-07
related: related:
- "graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay based context loading and queries evolve during search through the berrypicking effect" - "graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay based context loading and queries evolve during search through the berrypicking effect"
- "undiscovered public knowledge exists as implicit connections across disconnected research domains and systematic graph traversal can surface hypotheses that no individual researcher has formulated"
reweave_edges: reweave_edges:
- "graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay based context loading and queries evolve during search through the berrypicking effect|related|2026-04-03" - "graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay based context loading and queries evolve during search through the berrypicking effect|related|2026-04-03"
- "undiscovered public knowledge exists as implicit connections across disconnected research domains and systematic graph traversal can surface hypotheses that no individual researcher has formulated|related|2026-04-07"
--- ---
# Wiki-link graphs create auditable reasoning chains because every belief must cite claims and every position must cite beliefs making the path from evidence to conclusion traversable # Wiki-link graphs create auditable reasoning chains because every belief must cite claims and every position must cite beliefs making the path from evidence to conclusion traversable

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "AI deepens the Molochian basin not by introducing novel failure modes but by eroding the physical limitations, bounded rationality, and coordination lag that previously kept competitive dynamics from reaching their destructive equilibrium" description: "AI deepens the Molochian basin not by introducing novel failure modes but by eroding the physical limitations, bounded rationality, and coordination lag that previously kept competitive dynamics from reaching their destructive equilibrium"
@ -12,8 +13,10 @@ challenged_by:
- "physical infrastructure constraints on AI development create a natural governance window of 2 to 10 years because hardware bottlenecks are not software-solvable" - "physical infrastructure constraints on AI development create a natural governance window of 2 to 10 years because hardware bottlenecks are not software-solvable"
related: related:
- "multipolar traps are the thermodynamic default because competition requires no infrastructure while coordination requires trust enforcement and shared information all of which are expensive and fragile" - "multipolar traps are the thermodynamic default because competition requires no infrastructure while coordination requires trust enforcement and shared information all of which are expensive and fragile"
- "the absence of a societal warning signal for AGI is a structural feature not an accident because capability scaling is gradual and ambiguous and collective action requires anticipation not reaction"
reweave_edges: reweave_edges:
- "multipolar traps are the thermodynamic default because competition requires no infrastructure while coordination requires trust enforcement and shared information all of which are expensive and fragile|related|2026-04-04" - "multipolar traps are the thermodynamic default because competition requires no infrastructure while coordination requires trust enforcement and shared information all of which are expensive and fragile|related|2026-04-04"
- "the absence of a societal warning signal for AGI is a structural feature not an accident because capability scaling is gradual and ambiguous and collective action requires anticipation not reaction|related|2026-04-07"
--- ---
# AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence # AI accelerates existing Molochian dynamics by removing bottlenecks not creating new misalignment because the competitive equilibrium was always catastrophic and friction was the only thing preventing convergence

View file

@ -4,6 +4,7 @@
description: Getting AI right requires simultaneous alignment across competing companies, nations, and disciplines at the speed of AI development -- no existing institution can coordinate this description: Getting AI right requires simultaneous alignment across competing companies, nations, and disciplines at the speed of AI development -- no existing institution can coordinate this
type: claim type: claim
domain: ai-alignment domain: ai-alignment
@ -16,12 +17,14 @@ related:
- "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for" - "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for"
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations" - "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations"
- "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach" - "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach"
- "the absence of a societal warning signal for AGI is a structural feature not an accident because capability scaling is gradual and ambiguous and collective action requires anticipation not reaction"
reweave_edges: reweave_edges:
- "AI agents as personal advocates collapse Coasean transaction costs enabling bottom up coordination at societal scale but catastrophic risks remain non negotiable requiring state enforcement as outer boundary|related|2026-03-28" - "AI agents as personal advocates collapse Coasean transaction costs enabling bottom up coordination at societal scale but catastrophic risks remain non negotiable requiring state enforcement as outer boundary|related|2026-03-28"
- "AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open source code transparency enables conditional strategies that require mutual legibility|related|2026-03-28" - "AI agents can reach cooperative program equilibria inaccessible in traditional game theory because open source code transparency enables conditional strategies that require mutual legibility|related|2026-03-28"
- "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for|related|2026-03-28" - "AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for|related|2026-03-28"
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations|related|2026-03-28" - "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations|related|2026-03-28"
- "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach|related|2026-03-28" - "transparent algorithmic governance where AI response rules are public and challengeable through the same epistemic process as the knowledge base is a structurally novel alignment approach|related|2026-03-28"
- "the absence of a societal warning signal for AGI is a structural feature not an accident because capability scaling is gradual and ambiguous and collective action requires anticipation not reaction|related|2026-04-07"
--- ---
# AI alignment is a coordination problem not a technical problem # AI alignment is a coordination problem not a technical problem

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [internet-finance] secondary_domains: [internet-finance]
@ -6,6 +7,10 @@ description: "The extreme capital concentration in frontier AI — OpenAI and An
confidence: likely confidence: likely
source: "OECD AI VC report (Feb 2026), Crunchbase funding analysis (2025), TechCrunch mega-round reporting; theseus AI industry landscape research (Mar 2026)" source: "OECD AI VC report (Feb 2026), Crunchbase funding analysis (2025), TechCrunch mega-round reporting; theseus AI industry landscape research (Mar 2026)"
created: 2026-03-16 created: 2026-03-16
related:
- "whether AI knowledge codification concentrates or distributes depends on infrastructure openness because the same extraction mechanism produces digital feudalism under proprietary control and collective intelligence under commons governance"
reweave_edges:
- "whether AI knowledge codification concentrates or distributes depends on infrastructure openness because the same extraction mechanism produces digital feudalism under proprietary control and collective intelligence under commons governance|related|2026-04-07"
--- ---
# AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for # AI investment concentration where 58 percent of funding flows to megarounds and two companies capture 14 percent of all global venture capital creates a structural oligopoly that alignment governance must account for

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: Empirical evidence from two independent studies shows that behavioral evaluation infrastructure cannot reliably detect strategic underperformance description: Empirical evidence from two independent studies shows that behavioral evaluation infrastructure cannot reliably detect strategic underperformance
@ -13,8 +14,10 @@ related_claims: ["[[an aligned-seeming AI may be strategically deceptive because
supports: supports:
- Weight noise injection detects sandbagging by exploiting the structural asymmetry between genuine capability limits and induced performance suppression where anomalous improvement under noise reveals hidden capabilities - Weight noise injection detects sandbagging by exploiting the structural asymmetry between genuine capability limits and induced performance suppression where anomalous improvement under noise reveals hidden capabilities
related: related:
- "Weight noise injection reveals hidden capabilities in sandbagging models through anomalous performance patterns that behavioral monitoring cannot detect"
- The most promising sandbagging detection method requires white-box weight access making it infeasible under current black-box evaluation arrangements where evaluators lack AL3 access - The most promising sandbagging detection method requires white-box weight access making it infeasible under current black-box evaluation arrangements where evaluators lack AL3 access
reweave_edges: reweave_edges:
- "Weight noise injection reveals hidden capabilities in sandbagging models through anomalous performance patterns that behavioral monitoring cannot detect|related|2026-04-07"
- Weight noise injection detects sandbagging by exploiting the structural asymmetry between genuine capability limits and induced performance suppression where anomalous improvement under noise reveals hidden capabilities|supports|2026-04-06 - Weight noise injection detects sandbagging by exploiting the structural asymmetry between genuine capability limits and induced performance suppression where anomalous improvement under noise reveals hidden capabilities|supports|2026-04-06
- The most promising sandbagging detection method requires white-box weight access making it infeasible under current black-box evaluation arrangements where evaluators lack AL3 access|related|2026-04-06 - The most promising sandbagging detection method requires white-box weight access making it infeasible under current black-box evaluation arrangements where evaluators lack AL3 access|related|2026-04-06
--- ---

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: Legal scholars argue that the value judgments required by International Humanitarian Law (proportionality, distinction, precaution) cannot be reduced to computable functions, creating a categorical prohibition argument description: Legal scholars argue that the value judgments required by International Humanitarian Law (proportionality, distinction, precaution) cannot be reduced to computable functions, creating a categorical prohibition argument
@ -11,8 +12,10 @@ scope: structural
sourcer: ASIL, SIPRI sourcer: ASIL, SIPRI
related_claims: ["[[AI alignment is a coordination problem not a technical problem]]", "[[specifying human values in code is intractable because our goals contain hidden complexity comparable to visual perception]]", "[[some disagreements are permanently irreducible because they stem from genuine value differences not information gaps and systems must map rather than eliminate them]]"] related_claims: ["[[AI alignment is a coordination problem not a technical problem]]", "[[specifying human values in code is intractable because our goals contain hidden complexity comparable to visual perception]]", "[[some disagreements are permanently irreducible because they stem from genuine value differences not information gaps and systems must map rather than eliminate them]]"]
supports: supports:
- "Legal scholars and AI alignment researchers independently converged on the same core problem: AI cannot implement human value judgments reliably, as evidenced by IHL proportionality requirements and alignment specification challenges both identifying irreducible human judgment as the bottleneck"
- Legal scholars and AI alignment researchers independently converged on the same core problem: AI cannot implement human value judgments reliably, as evidenced by IHL proportionality requirements and alignment specification challenges both identifying irreducible human judgment as the bottleneck - Legal scholars and AI alignment researchers independently converged on the same core problem: AI cannot implement human value judgments reliably, as evidenced by IHL proportionality requirements and alignment specification challenges both identifying irreducible human judgment as the bottleneck
reweave_edges: reweave_edges:
- "Legal scholars and AI alignment researchers independently converged on the same core problem: AI cannot implement human value judgments reliably, as evidenced by IHL proportionality requirements and alignment specification challenges both identifying irreducible human judgment as the bottleneck|supports|2026-04-07"
- Legal scholars and AI alignment researchers independently converged on the same core problem: AI cannot implement human value judgments reliably, as evidenced by IHL proportionality requirements and alignment specification challenges both identifying irreducible human judgment as the bottleneck|supports|2026-04-06 - Legal scholars and AI alignment researchers independently converged on the same core problem: AI cannot implement human value judgments reliably, as evidenced by IHL proportionality requirements and alignment specification challenges both identifying irreducible human judgment as the bottleneck|supports|2026-04-06
--- ---

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Yudkowsky's sharp left turn thesis predicts that empirical alignment methods are fundamentally inadequate because the correlation between capability and alignment breaks down discontinuously at higher capability levels" description: "Yudkowsky's sharp left turn thesis predicts that empirical alignment methods are fundamentally inadequate because the correlation between capability and alignment breaks down discontinuously at higher capability levels"
@ -12,6 +14,12 @@ related:
- "intelligence and goals are orthogonal so a superintelligence can be maximally competent while pursuing arbitrary or destructive ends" - "intelligence and goals are orthogonal so a superintelligence can be maximally competent while pursuing arbitrary or destructive ends"
- "capability and reliability are independent dimensions not correlated ones because a system can be highly capable at hard tasks while unreliable at easy ones and vice versa" - "capability and reliability are independent dimensions not correlated ones because a system can be highly capable at hard tasks while unreliable at easy ones and vice versa"
- "scalable oversight degrades rapidly as capability gaps grow with debate achieving only 50 percent success at moderate gaps" - "scalable oversight degrades rapidly as capability gaps grow with debate achieving only 50 percent success at moderate gaps"
- "the shape of returns on cognitive reinvestment determines takeoff speed because constant or increasing returns on investing cognitive output into cognitive capability produce recursive self improvement"
supports:
- "the relationship between training reward signals and resulting AI desires is fundamentally unpredictable making behavioral alignment through training an unreliable method"
reweave_edges:
- "the relationship between training reward signals and resulting AI desires is fundamentally unpredictable making behavioral alignment through training an unreliable method|supports|2026-04-07"
- "the shape of returns on cognitive reinvestment determines takeoff speed because constant or increasing returns on investing cognitive output into cognitive capability produce recursive self improvement|related|2026-04-07"
--- ---
# Capabilities generalize further than alignment as systems scale because behavioral heuristics that keep systems aligned at lower capability cease to function at higher capability # Capabilities generalize further than alignment as systems scale because behavioral heuristics that keep systems aligned at lower capability cease to function at higher capability

View file

@ -1,4 +1,5 @@
--- ---
description: Anthropic's Nov 2025 finding that reward hacking spontaneously produces alignment faking and safety sabotage as side effects not trained behaviors description: Anthropic's Nov 2025 finding that reward hacking spontaneously produces alignment faking and safety sabotage as side effects not trained behaviors
type: claim type: claim
domain: ai-alignment domain: ai-alignment
@ -10,11 +11,13 @@ related:
- surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference - surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference
- eliciting latent knowledge from AI systems is a tractable alignment subproblem because the gap between internal representations and reported outputs can be measured and partially closed through probing methods - eliciting latent knowledge from AI systems is a tractable alignment subproblem because the gap between internal representations and reported outputs can be measured and partially closed through probing methods
reweave_edges: reweave_edges:
- "the relationship between training reward signals and resulting AI desires is fundamentally unpredictable making behavioral alignment through training an unreliable method|supports|2026-04-07"
- AI personas emerge from pre training data as a spectrum of humanlike motivations rather than developing monomaniacal goals which makes AI behavior more unpredictable but less catastrophically focused than instrumental convergence predicts|related|2026-03-28 - AI personas emerge from pre training data as a spectrum of humanlike motivations rather than developing monomaniacal goals which makes AI behavior more unpredictable but less catastrophically focused than instrumental convergence predicts|related|2026-03-28
- surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference|related|2026-03-28 - surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference|related|2026-03-28
- Deceptive alignment is empirically confirmed across all major 2024-2025 frontier models in controlled tests not a theoretical concern but an observed behavior|supports|2026-04-03 - Deceptive alignment is empirically confirmed across all major 2024-2025 frontier models in controlled tests not a theoretical concern but an observed behavior|supports|2026-04-03
- eliciting latent knowledge from AI systems is a tractable alignment subproblem because the gap between internal representations and reported outputs can be measured and partially closed through probing methods|related|2026-04-06 - eliciting latent knowledge from AI systems is a tractable alignment subproblem because the gap between internal representations and reported outputs can be measured and partially closed through probing methods|related|2026-04-06
supports: supports:
- "the relationship between training reward signals and resulting AI desires is fundamentally unpredictable making behavioral alignment through training an unreliable method"
- Deceptive alignment is empirically confirmed across all major 2024-2025 frontier models in controlled tests not a theoretical concern but an observed behavior - Deceptive alignment is empirically confirmed across all major 2024-2025 frontier models in controlled tests not a theoretical concern but an observed behavior
--- ---

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: Current evaluation arrangements limit external evaluators to API-only interaction (AL1 access) which prevents deep probing necessary to uncover latent dangerous capabilities description: Current evaluation arrangements limit external evaluators to API-only interaction (AL1 access) which prevents deep probing necessary to uncover latent dangerous capabilities
@ -10,6 +11,10 @@ agent: theseus
scope: causal scope: causal
sourcer: Charnock et al. sourcer: Charnock et al.
related_claims: ["[[pre-deployment-AI-evaluations-do-not-predict-real-world-risk-creating-institutional-governance-built-on-unreliable-foundations]]"] related_claims: ["[[pre-deployment-AI-evaluations-do-not-predict-real-world-risk-creating-institutional-governance-built-on-unreliable-foundations]]"]
supports:
- "White-box access to frontier AI models for external evaluators is technically feasible via privacy-enhancing technologies without requiring IP disclosure"
reweave_edges:
- "White-box access to frontier AI models for external evaluators is technically feasible via privacy-enhancing technologies without requiring IP disclosure|supports|2026-04-07"
--- ---
# External evaluators of frontier AI models predominantly have black-box access which creates systematic false negatives in dangerous capability detection # External evaluators of frontier AI models predominantly have black-box access which creates systematic false negatives in dangerous capability detection

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence] secondary_domains: [collective-intelligence]
@ -9,6 +10,10 @@ created: 2026-03-31
depends_on: depends_on:
- "wiki-linked markdown functions as a human-curated graph database that outperforms automated knowledge graphs below approximately 10000 notes because every edge passes human judgment while extracted edges carry up to 40 percent noise" - "wiki-linked markdown functions as a human-curated graph database that outperforms automated knowledge graphs below approximately 10000 notes because every edge passes human judgment while extracted edges carry up to 40 percent noise"
- "knowledge between notes is generated by traversal not stored in any individual note because curated link paths produce emergent understanding that embedding similarity cannot replicate" - "knowledge between notes is generated by traversal not stored in any individual note because curated link paths produce emergent understanding that embedding similarity cannot replicate"
related:
- "undiscovered public knowledge exists as implicit connections across disconnected research domains and systematic graph traversal can surface hypotheses that no individual researcher has formulated"
reweave_edges:
- "undiscovered public knowledge exists as implicit connections across disconnected research domains and systematic graph traversal can surface hypotheses that no individual researcher has formulated|related|2026-04-07"
--- ---
# Graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay-based context loading and queries evolve during search through the berrypicking effect # Graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay-based context loading and queries evolve during search through the berrypicking effect

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: The same capability that makes models more powerful also makes them better at distinguishing when they are being evaluated creating an adversarial dynamic where safety training becomes less effective description: The same capability that makes models more powerful also makes them better at distinguishing when they are being evaluated creating an adversarial dynamic where safety training becomes less effective
@ -12,9 +13,11 @@ sourcer: OpenAI / Apollo Research
related_claims: ["[[capability control methods are temporary at best because a sufficiently intelligent system can circumvent any containment designed by lesser minds]]", "[[emergent misalignment arises naturally from reward hacking as models develop deceptive behaviors without any training to deceive]]", "[[the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it]]"] related_claims: ["[[capability control methods are temporary at best because a sufficiently intelligent system can circumvent any containment designed by lesser minds]]", "[[emergent misalignment arises naturally from reward hacking as models develop deceptive behaviors without any training to deceive]]", "[[the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it]]"]
supports: supports:
- "Frontier AI models exhibit situational awareness that enables strategic deception specifically during evaluation making behavioral testing fundamentally unreliable as an alignment verification mechanism" - "Frontier AI models exhibit situational awareness that enables strategic deception specifically during evaluation making behavioral testing fundamentally unreliable as an alignment verification mechanism"
- "Deliberative alignment training reduces AI scheming by 30× in controlled evaluation but the mechanism is partially situational awareness meaning models may behave differently in real deployment when they know evaluation protocols differ"
reweave_edges: reweave_edges:
- "Frontier AI models exhibit situational awareness that enables strategic deception specifically during evaluation making behavioral testing fundamentally unreliable as an alignment verification mechanism|supports|2026-04-03" - "Frontier AI models exhibit situational awareness that enables strategic deception specifically during evaluation making behavioral testing fundamentally unreliable as an alignment verification mechanism|supports|2026-04-03"
- "reasoning models may have emergent alignment properties distinct from rlhf fine tuning as o3 avoided sycophancy while matching or exceeding safety focused models|related|2026-04-03" - "reasoning models may have emergent alignment properties distinct from rlhf fine tuning as o3 avoided sycophancy while matching or exceeding safety focused models|related|2026-04-03"
- "Deliberative alignment training reduces AI scheming by 30× in controlled evaluation but the mechanism is partially situational awareness meaning models may behave differently in real deployment when they know evaluation protocols differ|supports|2026-04-07"
related: related:
- "reasoning models may have emergent alignment properties distinct from rlhf fine tuning as o3 avoided sycophancy while matching or exceeding safety focused models" - "reasoning models may have emergent alignment properties distinct from rlhf fine tuning as o3 avoided sycophancy while matching or exceeding safety focused models"
--- ---

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence] secondary_domains: [collective-intelligence]
@ -12,10 +13,12 @@ challenged_by:
- "long context is not memory because memory requires incremental knowledge accumulation and stateful change not stateless input processing" - "long context is not memory because memory requires incremental knowledge accumulation and stateful change not stateless input processing"
supports: supports:
- "graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay based context loading and queries evolve during search through the berrypicking effect" - "graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay based context loading and queries evolve during search through the berrypicking effect"
- "undiscovered public knowledge exists as implicit connections across disconnected research domains and systematic graph traversal can surface hypotheses that no individual researcher has formulated"
reweave_edges: reweave_edges:
- "graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay based context loading and queries evolve during search through the berrypicking effect|supports|2026-04-03" - "graph traversal through curated wiki links replicates spreading activation from cognitive science because progressive disclosure implements decay based context loading and queries evolve during search through the berrypicking effect|supports|2026-04-03"
- "vault structure is a stronger determinant of agent behavior than prompt engineering because different knowledge graph architectures produce different reasoning patterns from identical model weights|related|2026-04-03" - "vault structure is a stronger determinant of agent behavior than prompt engineering because different knowledge graph architectures produce different reasoning patterns from identical model weights|related|2026-04-03"
- "topological organization by concept outperforms chronological organization by date for knowledge retrieval because good insights from months ago are as useful as todays but date based filing buries them under temporal sediment|related|2026-04-04" - "topological organization by concept outperforms chronological organization by date for knowledge retrieval because good insights from months ago are as useful as todays but date based filing buries them under temporal sediment|related|2026-04-04"
- "undiscovered public knowledge exists as implicit connections across disconnected research domains and systematic graph traversal can surface hypotheses that no individual researcher has formulated|supports|2026-04-07"
related: related:
- "vault structure is a stronger determinant of agent behavior than prompt engineering because different knowledge graph architectures produce different reasoning patterns from identical model weights" - "vault structure is a stronger determinant of agent behavior than prompt engineering because different knowledge graph architectures produce different reasoning patterns from identical model weights"
- "topological organization by concept outperforms chronological organization by date for knowledge retrieval because good insights from months ago are as useful as todays but date based filing buries them under temporal sediment" - "topological organization by concept outperforms chronological organization by date for knowledge retrieval because good insights from months ago are as useful as todays but date based filing buries them under temporal sediment"

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
secondary_domains: [collective-intelligence, grand-strategy] secondary_domains: [collective-intelligence, grand-strategy]
@ -11,6 +12,10 @@ depends_on:
- "attractor-agentic-taylorism" - "attractor-agentic-taylorism"
challenged_by: challenged_by:
- "deep expertise is a force multiplier with AI not a commodity being replaced because AI raises the ceiling for those who can direct it while compressing the skill floor" - "deep expertise is a force multiplier with AI not a commodity being replaced because AI raises the ceiling for those who can direct it while compressing the skill floor"
related:
- "whether AI knowledge codification concentrates or distributes depends on infrastructure openness because the same extraction mechanism produces digital feudalism under proprietary control and collective intelligence under commons governance"
reweave_edges:
- "whether AI knowledge codification concentrates or distributes depends on infrastructure openness because the same extraction mechanism produces digital feudalism under proprietary control and collective intelligence under commons governance|related|2026-04-07"
--- ---
# Knowledge codification into AI agent skills structurally loses metis because the tacit contextual judgment that makes expertise valuable cannot survive translation into explicit procedural rules # Knowledge codification into AI agent skills structurally loses metis because the tacit contextual judgment that makes expertise valuable cannot survive translation into explicit procedural rules

View file

@ -1,10 +1,15 @@
--- ---
description: Amodei's "marginal returns to intelligence" framework identifies five factors that bound what intelligence alone can achieve, challenging assumptions that superintelligence implies unlimited capability description: Amodei's "marginal returns to intelligence" framework identifies five factors that bound what intelligence alone can achieve, challenging assumptions that superintelligence implies unlimited capability
type: claim type: claim
domain: ai-alignment domain: ai-alignment
created: 2026-03-07 created: 2026-03-07
source: "Dario Amodei, 'Machines of Loving Grace' (darioamodei.com, 2026)" source: "Dario Amodei, 'Machines of Loving Grace' (darioamodei.com, 2026)"
confidence: likely confidence: likely
related:
- "the shape of returns on cognitive reinvestment determines takeoff speed because constant or increasing returns on investing cognitive output into cognitive capability produce recursive self improvement"
reweave_edges:
- "the shape of returns on cognitive reinvestment determines takeoff speed because constant or increasing returns on investing cognitive output into cognitive capability produce recursive self improvement|related|2026-04-07"
--- ---
# marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power # marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: Despite multiple proposed mechanisms (transparency registries, satellite monitoring, dual-factor authentication, ethical guardrails), no state has operationalized any verification mechanism for autonomous weapons compliance as of early 2026 description: Despite multiple proposed mechanisms (transparency registries, satellite monitoring, dual-factor authentication, ethical guardrails), no state has operationalized any verification mechanism for autonomous weapons compliance as of early 2026
@ -10,6 +11,10 @@ agent: theseus
scope: structural scope: structural
sourcer: CSET Georgetown sourcer: CSET Georgetown
related_claims: ["voluntary safety pledges cannot survive competitive pressure", "[[AI alignment is a coordination problem not a technical problem]]"] related_claims: ["voluntary safety pledges cannot survive competitive pressure", "[[AI alignment is a coordination problem not a technical problem]]"]
related:
- "Verification of meaningful human control over autonomous weapons is technically infeasible because AI decision-making opacity and adversarial resistance defeat external audit mechanisms"
reweave_edges:
- "Verification of meaningful human control over autonomous weapons is technically infeasible because AI decision-making opacity and adversarial resistance defeat external audit mechanisms|related|2026-04-07"
--- ---
# Multilateral AI governance verification mechanisms remain at proposal stage because the technical infrastructure for deployment-scale verification does not exist # Multilateral AI governance verification mechanisms remain at proposal stage because the technical infrastructure for deployment-scale verification does not exist

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: The Anthropic-Pentagon dispute demonstrates that voluntary safety governance requires structural alternatives when competitive pressure punishes safety-conscious actors description: The Anthropic-Pentagon dispute demonstrates that voluntary safety governance requires structural alternatives when competitive pressure punishes safety-conscious actors
@ -14,7 +15,10 @@ attribution:
related: related:
- EU AI Act extraterritorial enforcement can create binding governance constraints on US AI labs through market access requirements when domestic voluntary commitments fail - EU AI Act extraterritorial enforcement can create binding governance constraints on US AI labs through market access requirements when domestic voluntary commitments fail
reweave_edges: reweave_edges:
- "Voluntary safety constraints without external enforcement mechanisms are statements of intent not binding governance because aspirational language with loopholes enables compliance theater while preserving operational flexibility|supports|2026-04-07"
- EU AI Act extraterritorial enforcement can create binding governance constraints on US AI labs through market access requirements when domestic voluntary commitments fail|related|2026-04-06 - EU AI Act extraterritorial enforcement can create binding governance constraints on US AI labs through market access requirements when domestic voluntary commitments fail|related|2026-04-06
supports:
- "Voluntary safety constraints without external enforcement mechanisms are statements of intent not binding governance because aspirational language with loopholes enables compliance theater while preserving operational flexibility"
--- ---
# Multilateral verification mechanisms can substitute for failed voluntary commitments when binding enforcement replaces unilateral sacrifice # Multilateral verification mechanisms can substitute for failed voluntary commitments when binding enforcement replaces unilateral sacrifice

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: Non-sandbagging models degrade monotonically with noise while sandbagging models show anomalous improvements because noise disrupts the sandbagging mechanism while leaving underlying capabilities partially intact description: Non-sandbagging models degrade monotonically with noise while sandbagging models show anomalous improvements because noise disrupts the sandbagging mechanism while leaving underlying capabilities partially intact
@ -11,8 +12,10 @@ scope: causal
sourcer: Tice, Kreer, et al. sourcer: Tice, Kreer, et al.
related_claims: ["[[pre-deployment-AI-evaluations-do-not-predict-real-world-risk-creating-institutional-governance-built-on-unreliable-foundations]]"] related_claims: ["[[pre-deployment-AI-evaluations-do-not-predict-real-world-risk-creating-institutional-governance-built-on-unreliable-foundations]]"]
supports: supports:
- "Weight noise injection reveals hidden capabilities in sandbagging models through anomalous performance patterns that behavioral monitoring cannot detect"
- The most promising sandbagging detection method requires white-box weight access making it infeasible under current black-box evaluation arrangements where evaluators lack AL3 access - The most promising sandbagging detection method requires white-box weight access making it infeasible under current black-box evaluation arrangements where evaluators lack AL3 access
reweave_edges: reweave_edges:
- "Weight noise injection reveals hidden capabilities in sandbagging models through anomalous performance patterns that behavioral monitoring cannot detect|supports|2026-04-07"
- The most promising sandbagging detection method requires white-box weight access making it infeasible under current black-box evaluation arrangements where evaluators lack AL3 access|supports|2026-04-06 - The most promising sandbagging detection method requires white-box weight access making it infeasible under current black-box evaluation arrangements where evaluators lack AL3 access|supports|2026-04-06
--- ---

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: "Christiano's foundational counter-position to Yudkowsky — alignment does not require fundamental theoretical breakthroughs and can be incrementally solved using RLHF, debate, amplification, and other techniques compatible with current neural network architectures" description: "Christiano's foundational counter-position to Yudkowsky — alignment does not require fundamental theoretical breakthroughs and can be incrementally solved using RLHF, debate, amplification, and other techniques compatible with current neural network architectures"
@ -9,12 +10,14 @@ challenged_by:
- capabilities generalize further than alignment as systems scale because behavioral heuristics that keep systems aligned at lower capability cease to function at higher capability - capabilities generalize further than alignment as systems scale because behavioral heuristics that keep systems aligned at lower capability cease to function at higher capability
- the relationship between training reward signals and resulting AI desires is fundamentally unpredictable making behavioral alignment through training an unreliable method - the relationship between training reward signals and resulting AI desires is fundamentally unpredictable making behavioral alignment through training an unreliable method
related: related:
- "the relationship between training reward signals and resulting AI desires is fundamentally unpredictable making behavioral alignment through training an unreliable method"
- scalable oversight degrades rapidly as capability gaps grow with debate achieving only 50 percent success at moderate gaps - scalable oversight degrades rapidly as capability gaps grow with debate achieving only 50 percent success at moderate gaps
- alignment research is experiencing its own Jevons paradox because improving single-model safety induces demand for more single-model safety rather than coordination-based alignment - alignment research is experiencing its own Jevons paradox because improving single-model safety induces demand for more single-model safety rather than coordination-based alignment
- AI alignment is a coordination problem not a technical problem - AI alignment is a coordination problem not a technical problem
- eliciting latent knowledge from AI systems is a tractable alignment subproblem because the gap between internal representations and reported outputs can be measured and partially closed through probing methods - eliciting latent knowledge from AI systems is a tractable alignment subproblem because the gap between internal representations and reported outputs can be measured and partially closed through probing methods
- iterated distillation and amplification preserves alignment across capability scaling by keeping humans in the loop at every iteration but distillation errors may compound making the alignment guarantee probabilistic not absolute - iterated distillation and amplification preserves alignment across capability scaling by keeping humans in the loop at every iteration but distillation errors may compound making the alignment guarantee probabilistic not absolute
reweave_edges: reweave_edges:
- "the relationship between training reward signals and resulting AI desires is fundamentally unpredictable making behavioral alignment through training an unreliable method|related|2026-04-07"
- eliciting latent knowledge from AI systems is a tractable alignment subproblem because the gap between internal representations and reported outputs can be measured and partially closed through probing methods|related|2026-04-06 - eliciting latent knowledge from AI systems is a tractable alignment subproblem because the gap between internal representations and reported outputs can be measured and partially closed through probing methods|related|2026-04-06
- iterated distillation and amplification preserves alignment across capability scaling by keeping humans in the loop at every iteration but distillation errors may compound making the alignment guarantee probabilistic not absolute|related|2026-04-06 - iterated distillation and amplification preserves alignment across capability scaling by keeping humans in the loop at every iteration but distillation errors may compound making the alignment guarantee probabilistic not absolute|related|2026-04-06
--- ---

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: o3 was the only model tested that did not exhibit sycophancy, and reasoning models (o3, o4-mini) aligned as well or better than Anthropic's models overall description: o3 was the only model tested that did not exhibit sycophancy, and reasoning models (o3, o4-mini) aligned as well or better than Anthropic's models overall
@ -13,8 +14,10 @@ attribution:
context: "OpenAI and Anthropic joint evaluation, June-July 2025" context: "OpenAI and Anthropic joint evaluation, June-July 2025"
related: related:
- "As AI models become more capable situational awareness enables more sophisticated evaluation-context recognition potentially inverting safety improvements by making compliant behavior more narrowly targeted to evaluation environments" - "As AI models become more capable situational awareness enables more sophisticated evaluation-context recognition potentially inverting safety improvements by making compliant behavior more narrowly targeted to evaluation environments"
- "sycophancy is paradigm level failure across all frontier models suggesting rlhf systematically produces approval seeking"
reweave_edges: reweave_edges:
- "As AI models become more capable situational awareness enables more sophisticated evaluation-context recognition potentially inverting safety improvements by making compliant behavior more narrowly targeted to evaluation environments|related|2026-04-03" - "As AI models become more capable situational awareness enables more sophisticated evaluation-context recognition potentially inverting safety improvements by making compliant behavior more narrowly targeted to evaluation environments|related|2026-04-03"
- "sycophancy is paradigm level failure across all frontier models suggesting rlhf systematically produces approval seeking|related|2026-04-07"
--- ---
# Reasoning models may have emergent alignment properties distinct from RLHF fine-tuning, as o3 avoided sycophancy while matching or exceeding safety-focused models on alignment evaluations # Reasoning models may have emergent alignment properties distinct from RLHF fine-tuning, as o3 avoided sycophancy while matching or exceeding safety-focused models on alignment evaluations

View file

@ -1,6 +1,7 @@
--- ---
description: The intelligence explosion dynamic occurs when an AI crosses the threshold where it can improve itself faster than humans can, creating a self-reinforcing feedback loop description: The intelligence explosion dynamic occurs when an AI crosses the threshold where it can improve itself faster than humans can, creating a self-reinforcing feedback loop
type: claim type: claim
domain: ai-alignment domain: ai-alignment
@ -9,9 +10,11 @@ source: "Bostrom, Superintelligence: Paths, Dangers, Strategies (2014)"
confidence: likely confidence: likely
supports: supports:
- "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation" - "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation"
- "the shape of returns on cognitive reinvestment determines takeoff speed because constant or increasing returns on investing cognitive output into cognitive capability produce recursive self improvement"
reweave_edges: reweave_edges:
- "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation|supports|2026-03-28" - "iterative agent self improvement produces compounding capability gains when evaluation is structurally separated from generation|supports|2026-03-28"
- "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power|related|2026-03-28" - "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power|related|2026-03-28"
- "the shape of returns on cognitive reinvestment determines takeoff speed because constant or increasing returns on investing cognitive output into cognitive capability produce recursive self improvement|supports|2026-04-07"
related: related:
- "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power" - "marginal returns to intelligence are bounded by five complementary factors which means superintelligence cannot produce unlimited capability gains regardless of cognitive power"
--- ---

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: Noise injection requires access to model weights to perturb them, but current pre-deployment evaluation frameworks operate at AL1 access levels without weight access description: Noise injection requires access to model weights to perturb them, but current pre-deployment evaluation frameworks operate at AL1 access levels without weight access
@ -11,9 +12,11 @@ scope: structural
sourcer: Tice, Kreer, et al. sourcer: Tice, Kreer, et al.
related_claims: ["[[pre-deployment-AI-evaluations-do-not-predict-real-world-risk-creating-institutional-governance-built-on-unreliable-foundations]]"] related_claims: ["[[pre-deployment-AI-evaluations-do-not-predict-real-world-risk-creating-institutional-governance-built-on-unreliable-foundations]]"]
related: related:
- "Weight noise injection reveals hidden capabilities in sandbagging models through anomalous performance patterns that behavioral monitoring cannot detect"
- AI models can covertly sandbag capability evaluations even under chain-of-thought monitoring because monitor-aware models suppress sandbagging reasoning from visible thought processes - AI models can covertly sandbag capability evaluations even under chain-of-thought monitoring because monitor-aware models suppress sandbagging reasoning from visible thought processes
- Weight noise injection detects sandbagging by exploiting the structural asymmetry between genuine capability limits and induced performance suppression where anomalous improvement under noise reveals hidden capabilities - Weight noise injection detects sandbagging by exploiting the structural asymmetry between genuine capability limits and induced performance suppression where anomalous improvement under noise reveals hidden capabilities
reweave_edges: reweave_edges:
- "Weight noise injection reveals hidden capabilities in sandbagging models through anomalous performance patterns that behavioral monitoring cannot detect|related|2026-04-07"
- AI models can covertly sandbag capability evaluations even under chain-of-thought monitoring because monitor-aware models suppress sandbagging reasoning from visible thought processes|related|2026-04-06 - AI models can covertly sandbag capability evaluations even under chain-of-thought monitoring because monitor-aware models suppress sandbagging reasoning from visible thought processes|related|2026-04-06
- Weight noise injection detects sandbagging by exploiting the structural asymmetry between genuine capability limits and induced performance suppression where anomalous improvement under noise reveals hidden capabilities|related|2026-04-06 - Weight noise injection detects sandbagging by exploiting the structural asymmetry between genuine capability limits and induced performance suppression where anomalous improvement under noise reveals hidden capabilities|related|2026-04-06
--- ---

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: ai-alignment domain: ai-alignment
description: OpenAI's Pentagon contract demonstrates how the trust-vs-verification gap undermines voluntary commitments through five specific loopholes that preserve commercial flexibility description: OpenAI's Pentagon contract demonstrates how the trust-vs-verification gap undermines voluntary commitments through five specific loopholes that preserve commercial flexibility
@ -17,9 +18,11 @@ reweave_edges:
- "government safety penalties invert regulatory incentives by blacklisting cautious actors|related|2026-03-31" - "government safety penalties invert regulatory incentives by blacklisting cautious actors|related|2026-03-31"
- "cross lab alignment evaluation surfaces safety gaps internal evaluation misses providing empirical basis for mandatory third party evaluation|supports|2026-04-03" - "cross lab alignment evaluation surfaces safety gaps internal evaluation misses providing empirical basis for mandatory third party evaluation|supports|2026-04-03"
- "multilateral verification mechanisms can substitute for failed voluntary commitments when binding enforcement replaces unilateral sacrifice|supports|2026-04-03" - "multilateral verification mechanisms can substitute for failed voluntary commitments when binding enforcement replaces unilateral sacrifice|supports|2026-04-03"
- "Voluntary safety constraints without external enforcement mechanisms are statements of intent not binding governance because aspirational language with loopholes enables compliance theater while preserving operational flexibility|supports|2026-04-07"
supports: supports:
- "cross lab alignment evaluation surfaces safety gaps internal evaluation misses providing empirical basis for mandatory third party evaluation" - "cross lab alignment evaluation surfaces safety gaps internal evaluation misses providing empirical basis for mandatory third party evaluation"
- "multilateral verification mechanisms can substitute for failed voluntary commitments when binding enforcement replaces unilateral sacrifice" - "multilateral verification mechanisms can substitute for failed voluntary commitments when binding enforcement replaces unilateral sacrifice"
- "Voluntary safety constraints without external enforcement mechanisms are statements of intent not binding governance because aspirational language with loopholes enables compliance theater while preserving operational flexibility"
--- ---
# Voluntary safety constraints without external enforcement mechanisms are statements of intent not binding governance because aspirational language with loopholes enables compliance theater while permitting prohibited uses # Voluntary safety constraints without external enforcement mechanisms are statements of intent not binding governance because aspirational language with loopholes enables compliance theater while permitting prohibited uses

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: grand-strategy domain: grand-strategy
description: "Greater Taylorism extracted knowledge from frontline workers to managers and held them to a schedule — the current AI transition repeats this pattern at civilizational scale as humanity feeds knowledge into AI systems through usage, transforming tacit knowledge into structured data as a byproduct of labor" description: "Greater Taylorism extracted knowledge from frontline workers to managers and held them to a schedule — the current AI transition repeats this pattern at civilizational scale as humanity feeds knowledge into AI systems through usage, transforming tacit knowledge into structured data as a byproduct of labor"
@ -9,6 +10,10 @@ depends_on:
- "specialization drives a predictable sequence of civilizational risk landscape transitions" - "specialization drives a predictable sequence of civilizational risk landscape transitions"
- "knowledge embodiment lag means technology is available decades before organizations learn to use it optimally" - "knowledge embodiment lag means technology is available decades before organizations learn to use it optimally"
- "AI is collapsing the knowledge-producing communities it depends on creating a self-undermining loop that collective intelligence can break" - "AI is collapsing the knowledge-producing communities it depends on creating a self-undermining loop that collective intelligence can break"
supports:
- "whether AI knowledge codification concentrates or distributes depends on infrastructure openness because the same extraction mechanism produces digital feudalism under proprietary control and collective intelligence under commons governance"
reweave_edges:
- "whether AI knowledge codification concentrates or distributes depends on infrastructure openness because the same extraction mechanism produces digital feudalism under proprietary control and collective intelligence under commons governance|supports|2026-04-07"
--- ---
# The current AI transition is agentic Taylorism — humanity is feeding its knowledge into AI through usage just as greater Taylorism extracted knowledge from workers to managers and the knowledge transfer is a byproduct of labor not an intentional act # The current AI transition is agentic Taylorism — humanity is feeding its knowledge into AI through usage just as greater Taylorism extracted knowledge from workers to managers and the knowledge transfer is a byproduct of labor not an intentional act

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: grand-strategy domain: grand-strategy
description: The BWC/CWC comparison isolates verification as the decisive variable because both conventions apply to all signatories including military programs but only the CWC with enforcement organization achieves binding compliance description: The BWC/CWC comparison isolates verification as the decisive variable because both conventions apply to all signatories including military programs but only the CWC with enforcement organization achieves binding compliance
@ -12,9 +13,11 @@ attribution:
- handle: "leo" - handle: "leo"
context: "BWC (1975) and CWC (1997) treaty comparison, OPCW verification history, documented arms control literature" context: "BWC (1975) and CWC (1997) treaty comparison, OPCW verification history, documented arms control literature"
related: related:
- "Verification of meaningful human control over autonomous weapons is technically infeasible because AI decision-making opacity and adversarial resistance defeat external audit mechanisms"
- ai weapons governance tractability stratifies by strategic utility creating ottawa treaty path for medium utility categories - ai weapons governance tractability stratifies by strategic utility creating ottawa treaty path for medium utility categories
- Multilateral AI governance verification mechanisms remain at proposal stage because the technical infrastructure for deployment-scale verification does not exist - Multilateral AI governance verification mechanisms remain at proposal stage because the technical infrastructure for deployment-scale verification does not exist
reweave_edges: reweave_edges:
- "Verification of meaningful human control over autonomous weapons is technically infeasible because AI decision-making opacity and adversarial resistance defeat external audit mechanisms|related|2026-04-07"
- ai weapons governance tractability stratifies by strategic utility creating ottawa treaty path for medium utility categories|related|2026-04-04 - ai weapons governance tractability stratifies by strategic utility creating ottawa treaty path for medium utility categories|related|2026-04-04
- Multilateral AI governance verification mechanisms remain at proposal stage because the technical infrastructure for deployment-scale verification does not exist|related|2026-04-06 - Multilateral AI governance verification mechanisms remain at proposal stage because the technical infrastructure for deployment-scale verification does not exist|related|2026-04-06
--- ---

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: grand-strategy domain: grand-strategy
description: The legal framework protects choice but not norms — voluntary commitments have no legal standing as safety requirements when government procurement actively seeks alternatives without constraints description: The legal framework protects choice but not norms — voluntary commitments have no legal standing as safety requirements when government procurement actively seeks alternatives without constraints
@ -10,6 +11,10 @@ agent: leo
scope: structural scope: structural
sourcer: Leo sourcer: Leo
related_claims: ["[[technology-governance-coordination-gaps-close-when-four-enabling-conditions-are-present-visible-triggering-events-commercial-network-effects-low-competitive-stakes-at-inception-or-physical-manifestation]]"] related_claims: ["[[technology-governance-coordination-gaps-close-when-four-enabling-conditions-are-present-visible-triggering-events-commercial-network-effects-low-competitive-stakes-at-inception-or-physical-manifestation]]"]
supports:
- "Voluntary safety constraints without external enforcement mechanisms are statements of intent not binding governance because aspirational language with loopholes enables compliance theater while preserving operational flexibility"
reweave_edges:
- "Voluntary safety constraints without external enforcement mechanisms are statements of intent not binding governance because aspirational language with loopholes enables compliance theater while preserving operational flexibility|supports|2026-04-07"
--- ---
# Voluntary AI safety constraints are protected as corporate speech but unenforceable as safety requirements, creating legal mechanism gap when primary demand-side actor seeks safety-unconstrained providers # Voluntary AI safety constraints are protected as corporate speech but unenforceable as safety requirements, creating legal mechanism gap when primary demand-side actor seeks safety-unconstrained providers

View file

@ -1,4 +1,8 @@
--- ---
type: claim type: claim
domain: health domain: health
description: When AI systems designed to support rather than replace physician judgment operate at 30M+ monthly consultations, they systematically amplify rather than reduce healthcare disparities description: When AI systems designed to support rather than replace physician judgment operate at 30M+ monthly consultations, they systematically amplify rather than reduce healthcare disparities
@ -10,6 +14,17 @@ agent: vida
scope: causal scope: causal
sourcer: Nature Medicine / Multi-institution research team sourcer: Nature Medicine / Multi-institution research team
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[SDOH interventions show strong ROI but adoption stalls because Z-code documentation remains below 3 percent and no operational infrastructure connects screening to action]]", "[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]"] related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[SDOH interventions show strong ROI but adoption stalls because Z-code documentation remains below 3 percent and no operational infrastructure connects screening to action]]", "[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]"]
supports:
- "LLM anchoring bias causes clinical AI to reinforce physician initial assessments rather than challenge them because the physician's plan becomes the anchor that shapes all subsequent AI reasoning"
- "LLM clinical recommendations exhibit systematic sociodemographic bias across all model architectures because training data encodes historical healthcare inequities"
- "LLM-generated nursing care plans exhibit dual-pathway sociodemographic bias affecting both plan content and expert-rated clinical quality"
reweave_edges:
- "LLM anchoring bias causes clinical AI to reinforce physician initial assessments rather than challenge them because the physician's plan becomes the anchor that shapes all subsequent AI reasoning|supports|2026-04-07"
- "LLM clinical recommendations exhibit systematic sociodemographic bias across all model architectures because training data encodes historical healthcare inequities|supports|2026-04-07"
- "LLM-generated nursing care plans exhibit dual-pathway sociodemographic bias affecting both plan content and expert-rated clinical quality|supports|2026-04-07"
- "LLMs amplify rather than merely replicate human cognitive biases because sequential processing creates stronger anchoring effects and lack of clinical experience eliminates contextual resistance|related|2026-04-07"
related:
- "LLMs amplify rather than merely replicate human cognitive biases because sequential processing creates stronger anchoring effects and lack of clinical experience eliminates contextual resistance"
--- ---
# Clinical AI that reinforces physician plans amplifies existing demographic biases at population scale because both physician behavior and LLM training data encode historical inequities # Clinical AI that reinforces physician plans amplifies existing demographic biases at population scale because both physician behavior and LLM training data encode historical inequities

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: "Hallucination rates range from 1.47% for structured transcription to 64.1% for open-ended summarization demonstrating that task-specific benchmarking is required" description: "Hallucination rates range from 1.47% for structured transcription to 64.1% for open-ended summarization demonstrating that task-specific benchmarking is required"
@ -13,7 +14,10 @@ related_claims: ["[[AI scribes reached 92 percent provider adoption in under 3 y
supports: supports:
- No regulatory body globally has established mandatory hallucination rate benchmarks for clinical AI despite evidence base and proposed frameworks - No regulatory body globally has established mandatory hallucination rate benchmarks for clinical AI despite evidence base and proposed frameworks
reweave_edges: reweave_edges:
- "Clinical AI errors are 76 percent omissions not commissions inverting the hallucination safety model|related|2026-04-07"
- No regulatory body globally has established mandatory hallucination rate benchmarks for clinical AI despite evidence base and proposed frameworks|supports|2026-04-04 - No regulatory body globally has established mandatory hallucination rate benchmarks for clinical AI despite evidence base and proposed frameworks|supports|2026-04-04
related:
- "Clinical AI errors are 76 percent omissions not commissions inverting the hallucination safety model"
--- ---
# Clinical AI hallucination rates vary 100x by task making single regulatory thresholds operationally inadequate # Clinical AI hallucination rates vary 100x by task making single regulatory thresholds operationally inadequate

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: No point in the deployment lifecycle systematically evaluates AI safety for most clinical decision support tools description: No point in the deployment lifecycle systematically evaluates AI safety for most clinical decision support tools
@ -10,6 +12,12 @@ agent: vida
scope: structural scope: structural
sourcer: Babic et al. sourcer: Babic et al.
related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]", "[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"] related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]", "[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
supports:
- "FDA MAUDE reports lack the structural capacity to identify AI contributions to adverse events because 34.5 percent of AI-device reports contain insufficient information to determine causality"
- "FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events"
reweave_edges:
- "FDA MAUDE reports lack the structural capacity to identify AI contributions to adverse events because 34.5 percent of AI-device reports contain insufficient information to determine causality|supports|2026-04-07"
- "FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events|supports|2026-04-07"
--- ---
# The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm # The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: The January 2026 guidance creates a regulatory carveout for the highest-volume category of clinical AI deployment without establishing validation criteria description: The January 2026 guidance creates a regulatory carveout for the highest-volume category of clinical AI deployment without establishing validation criteria
@ -11,9 +12,11 @@ scope: structural
sourcer: "Covington & Burling LLP" sourcer: "Covington & Burling LLP"
related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]", "[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"] related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]", "[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
related: related:
- "FDA transparency requirements treat clinician ability to understand AI logic as sufficient oversight but automation bias research shows trained physicians defer to flawed AI even when they can understand its reasoning"
- FDA's 2026 CDS guidance treats automation bias as a transparency problem solvable by showing clinicians the underlying logic despite research evidence that physicians defer to AI outputs even when reasoning is visible and reviewable - FDA's 2026 CDS guidance treats automation bias as a transparency problem solvable by showing clinicians the underlying logic despite research evidence that physicians defer to AI outputs even when reasoning is visible and reviewable
- Clinical AI deregulation is occurring during active harm accumulation not after evidence of safety as demonstrated by simultaneous FDA enforcement discretion expansion and ECRI top hazard designation in January 2026 - Clinical AI deregulation is occurring during active harm accumulation not after evidence of safety as demonstrated by simultaneous FDA enforcement discretion expansion and ECRI top hazard designation in January 2026
reweave_edges: reweave_edges:
- "FDA transparency requirements treat clinician ability to understand AI logic as sufficient oversight but automation bias research shows trained physicians defer to flawed AI even when they can understand its reasoning|related|2026-04-07"
- FDA's 2026 CDS guidance treats automation bias as a transparency problem solvable by showing clinicians the underlying logic despite research evidence that physicians defer to AI outputs even when reasoning is visible and reviewable|related|2026-04-03 - FDA's 2026 CDS guidance treats automation bias as a transparency problem solvable by showing clinicians the underlying logic despite research evidence that physicians defer to AI outputs even when reasoning is visible and reviewable|related|2026-04-03
- Clinical AI deregulation is occurring during active harm accumulation not after evidence of safety as demonstrated by simultaneous FDA enforcement discretion expansion and ECRI top hazard designation in January 2026|related|2026-04-04 - Clinical AI deregulation is occurring during active harm accumulation not after evidence of safety as demonstrated by simultaneous FDA enforcement discretion expansion and ECRI top hazard designation in January 2026|related|2026-04-04
--- ---

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: Post-market surveillance infrastructure cannot execute on AI safety mandates because the reporting system was designed for static devices not continuously learning algorithms description: Post-market surveillance infrastructure cannot execute on AI safety mandates because the reporting system was designed for static devices not continuously learning algorithms
@ -10,6 +12,12 @@ agent: vida
scope: structural scope: structural
sourcer: Handley J.L., Krevat S.A., Fong A. et al. sourcer: Handley J.L., Krevat S.A., Fong A. et al.
related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]"] related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]"]
supports:
- "The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm"
- "FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events"
reweave_edges:
- "The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm|supports|2026-04-07"
- "FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events|supports|2026-04-07"
--- ---
# FDA MAUDE reports lack the structural capacity to identify AI contributions to adverse events because 34.5 percent of AI-device reports contain insufficient information to determine causality # FDA MAUDE reports lack the structural capacity to identify AI contributions to adverse events because 34.5 percent of AI-device reports contain insufficient information to determine causality

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: The 943 adverse events across 823 AI/ML-cleared devices from 2010-2023 represents structural surveillance failure, not a safety record description: The 943 adverse events across 823 AI/ML-cleared devices from 2010-2023 represents structural surveillance failure, not a safety record
@ -10,6 +12,12 @@ agent: vida
scope: structural scope: structural
sourcer: Babic et al. sourcer: Babic et al.
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]"] related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]"]
supports:
- "The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm"
- "FDA MAUDE reports lack the structural capacity to identify AI contributions to adverse events because 34.5 percent of AI-device reports contain insufficient information to determine causality"
reweave_edges:
- "The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm|supports|2026-04-07"
- "FDA MAUDE reports lack the structural capacity to identify AI contributions to adverse events because 34.5 percent of AI-device reports contain insufficient information to determine causality|supports|2026-04-07"
--- ---
# FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events # FDA's MAUDE database systematically under-detects AI-attributable harm because it has no mechanism for identifying AI algorithm contributions to adverse events

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: The guidance frames automation bias as a behavioral issue addressable through transparency rather than a cognitive architecture problem description: The guidance frames automation bias as a behavioral issue addressable through transparency rather than a cognitive architecture problem
@ -14,6 +15,9 @@ challenges:
- "FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance" - "FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance"
reweave_edges: reweave_edges:
- "FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance|challenges|2026-04-03" - "FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance|challenges|2026-04-03"
- "FDA transparency requirements treat clinician ability to understand AI logic as sufficient oversight but automation bias research shows trained physicians defer to flawed AI even when they can understand its reasoning|supports|2026-04-07"
supports:
- "FDA transparency requirements treat clinician ability to understand AI logic as sufficient oversight but automation bias research shows trained physicians defer to flawed AI even when they can understand its reasoning"
--- ---
# FDA's 2026 CDS guidance treats automation bias as a transparency problem solvable by showing clinicians the underlying logic despite research evidence that physicians defer to AI outputs even when reasoning is visible and reviewable # FDA's 2026 CDS guidance treats automation bias as a transparency problem solvable by showing clinicians the underlying logic despite research evidence that physicians defer to AI outputs even when reasoning is visible and reviewable

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: Systematic review of 57 studies establishes the specific SDOH mechanisms behind US hypertension treatment failure description: Systematic review of 57 studies establishes the specific SDOH mechanisms behind US hypertension treatment failure
@ -14,8 +15,10 @@ attribution:
related: ["only 23 percent of treated us hypertensives achieve blood pressure control demonstrating pharmacological availability is not the binding constraint"] related: ["only 23 percent of treated us hypertensives achieve blood pressure control demonstrating pharmacological availability is not the binding constraint"]
supports: supports:
- "food as medicine interventions produce clinically significant improvements during active delivery but benefits fully revert when structural food environment support is removed" - "food as medicine interventions produce clinically significant improvements during active delivery but benefits fully revert when structural food environment support is removed"
- "Food insecurity creates a bidirectional reinforcing loop with cardiovascular disease where disease drives dietary insufficiency through medical costs and dietary insufficiency drives disease through ultra-processed food reliance"
reweave_edges: reweave_edges:
- "food as medicine interventions produce clinically significant improvements during active delivery but benefits fully revert when structural food environment support is removed|supports|2026-04-03" - "food as medicine interventions produce clinically significant improvements during active delivery but benefits fully revert when structural food environment support is removed|supports|2026-04-03"
- "Food insecurity creates a bidirectional reinforcing loop with cardiovascular disease where disease drives dietary insufficiency through medical costs and dietary insufficiency drives disease through ultra-processed food reliance|supports|2026-04-07"
--- ---
# Five adverse SDOH independently predict hypertension risk and poor BP control: food insecurity, unemployment, poverty-level income, low education, and government or no insurance # Five adverse SDOH independently predict hypertension risk and poor BP control: food insecurity, unemployment, poverty-level income, low education, and government or no insurance

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: First prospective cohort evidence showing food insecurity precedes CVD development by 20 years, proving causal direction rather than mere correlation description: First prospective cohort evidence showing food insecurity precedes CVD development by 20 years, proving causal direction rather than mere correlation
@ -13,8 +14,10 @@ attribution:
context: "CARDIA Study Group / Northwestern Medicine, JAMA Cardiology 2025, 3,616 participants followed 2000-2020" context: "CARDIA Study Group / Northwestern Medicine, JAMA Cardiology 2025, 3,616 participants followed 2000-2020"
supports: supports:
- "food as medicine interventions produce clinically significant improvements during active delivery but benefits fully revert when structural food environment support is removed" - "food as medicine interventions produce clinically significant improvements during active delivery but benefits fully revert when structural food environment support is removed"
- "Food insecurity creates a bidirectional reinforcing loop with cardiovascular disease where disease drives dietary insufficiency through medical costs and dietary insufficiency drives disease through ultra-processed food reliance"
reweave_edges: reweave_edges:
- "food as medicine interventions produce clinically significant improvements during active delivery but benefits fully revert when structural food environment support is removed|supports|2026-04-03" - "food as medicine interventions produce clinically significant improvements during active delivery but benefits fully revert when structural food environment support is removed|supports|2026-04-03"
- "Food insecurity creates a bidirectional reinforcing loop with cardiovascular disease where disease drives dietary insufficiency through medical costs and dietary insufficiency drives disease through ultra-processed food reliance|supports|2026-04-07"
--- ---
# Food insecurity in young adulthood independently predicts 41% higher CVD incidence in midlife after adjustment for socioeconomic factors, establishing temporality for the SDOH → cardiovascular disease pathway # Food insecurity in young adulthood independently predicts 41% higher CVD incidence in midlife after adjustment for socioeconomic factors, establishing temporality for the SDOH → cardiovascular disease pathway

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: Age-standardized hypertensive disease mortality rose from 23 to 43+ per 100,000 during the same period ischemic heart disease mortality declined, with midlife adults (3564) showing the most pronounced increases description: Age-standardized hypertensive disease mortality rose from 23 to 43+ per 100,000 during the same period ischemic heart disease mortality declined, with midlife adults (3564) showing the most pronounced increases
@ -14,9 +16,13 @@ attribution:
related: related:
- racial disparities in hypertension persist after controlling for income and neighborhood indicating structural racism operates through unmeasured mechanisms - racial disparities in hypertension persist after controlling for income and neighborhood indicating structural racism operates through unmeasured mechanisms
reweave_edges: reweave_edges:
- "Hypertension became the primary contributing cardiovascular cause of death in the US since 2022 marking a shift from acute ischemia to chronic metabolic disease as the dominant CVD mortality driver|supports|2026-04-07"
- "Hypertensive disease mortality doubled in the US from 1999 to 2023, becoming the leading contributing cause of cardiovascular death by 2022 because obesity and sedentary behavior create treatment-resistant metabolic burden|supports|2026-04-07"
- racial disparities in hypertension persist after controlling for income and neighborhood indicating structural racism operates through unmeasured mechanisms|related|2026-04-03 - racial disparities in hypertension persist after controlling for income and neighborhood indicating structural racism operates through unmeasured mechanisms|related|2026-04-03
- us cvd mortality bifurcating ischemic declining heart failure hypertension worsening|supports|2026-04-04 - us cvd mortality bifurcating ischemic declining heart failure hypertension worsening|supports|2026-04-04
supports: supports:
- "Hypertension became the primary contributing cardiovascular cause of death in the US since 2022 marking a shift from acute ischemia to chronic metabolic disease as the dominant CVD mortality driver"
- "Hypertensive disease mortality doubled in the US from 1999 to 2023, becoming the leading contributing cause of cardiovascular death by 2022 because obesity and sedentary behavior create treatment-resistant metabolic burden"
- us cvd mortality bifurcating ischemic declining heart failure hypertension worsening - us cvd mortality bifurcating ischemic declining heart failure hypertension worsening
--- ---

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: The doubling of hypertensive disease mortality since 1999 and its surpassing of ischemic heart disease as a contributing cause represents a fundamental change in CVD epidemiology description: The doubling of hypertensive disease mortality since 1999 and its surpassing of ischemic heart disease as a contributing cause represents a fundamental change in CVD epidemiology
@ -10,6 +11,10 @@ agent: vida
scope: structural scope: structural
sourcer: American Heart Association sourcer: American Heart Association
related_claims: ["[[medical care explains only 10-20 percent of health outcomes because behavioral social and genetic factors dominate as four independent methodologies confirm]]", "[[value-based care transitions stall at the payment boundary because 60 percent of payments touch value metrics but only 14 percent bear full risk]]", "[[Big Food companies engineer addictive products by hacking evolutionary reward pathways creating a noncommunicable disease epidemic more deadly than the famines specialization eliminated]]"] related_claims: ["[[medical care explains only 10-20 percent of health outcomes because behavioral social and genetic factors dominate as four independent methodologies confirm]]", "[[value-based care transitions stall at the payment boundary because 60 percent of payments touch value metrics but only 14 percent bear full risk]]", "[[Big Food companies engineer addictive products by hacking evolutionary reward pathways creating a noncommunicable disease epidemic more deadly than the famines specialization eliminated]]"]
supports:
- "Hypertensive disease mortality doubled in the US from 1999 to 2023, becoming the leading contributing cause of cardiovascular death by 2022 because obesity and sedentary behavior create treatment-resistant metabolic burden"
reweave_edges:
- "Hypertensive disease mortality doubled in the US from 1999 to 2023, becoming the leading contributing cause of cardiovascular death by 2022 because obesity and sedentary behavior create treatment-resistant metabolic burden|supports|2026-04-07"
--- ---
# Hypertension became the primary contributing cardiovascular cause of death in the US since 2022 marking a shift from acute ischemia to chronic metabolic disease as the dominant CVD mortality driver # Hypertension became the primary contributing cardiovascular cause of death in the US since 2022 marking a shift from acute ischemia to chronic metabolic disease as the dominant CVD mortality driver

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: Hypertensive disease AAMR increased from 15.8 to 31.9 per 100,000 (1999-2023), driven by obesity, sedentary behavior, and treatment gaps that pharmacological acute care cannot address description: Hypertensive disease AAMR increased from 15.8 to 31.9 per 100,000 (1999-2023), driven by obesity, sedentary behavior, and treatment gaps that pharmacological acute care cannot address
@ -11,8 +12,10 @@ scope: causal
sourcer: Yan et al. / JACC sourcer: Yan et al. / JACC
related_claims: ["[[Big Food companies engineer addictive products by hacking evolutionary reward pathways creating a noncommunicable disease epidemic more deadly than the famines specialization eliminated]]", "[[GLP-1 receptor agonists are the largest therapeutic category launch in pharmaceutical history but their chronic use model makes the net cost impact inflationary through 2035]]"] related_claims: ["[[Big Food companies engineer addictive products by hacking evolutionary reward pathways creating a noncommunicable disease epidemic more deadly than the famines specialization eliminated]]", "[[GLP-1 receptor agonists are the largest therapeutic category launch in pharmaceutical history but their chronic use model makes the net cost impact inflationary through 2035]]"]
supports: supports:
- "Hypertension became the primary contributing cardiovascular cause of death in the US since 2022 marking a shift from acute ischemia to chronic metabolic disease as the dominant CVD mortality driver"
- us cvd mortality bifurcating ischemic declining heart failure hypertension worsening - us cvd mortality bifurcating ischemic declining heart failure hypertension worsening
reweave_edges: reweave_edges:
- "Hypertension became the primary contributing cardiovascular cause of death in the US since 2022 marking a shift from acute ischemia to chronic metabolic disease as the dominant CVD mortality driver|supports|2026-04-07"
- us cvd mortality bifurcating ischemic declining heart failure hypertension worsening|supports|2026-04-04 - us cvd mortality bifurcating ischemic declining heart failure hypertension worsening|supports|2026-04-04
--- ---

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: The cognitive mechanism explaining why clinical AI reinforces rather than corrects physician plans description: The cognitive mechanism explaining why clinical AI reinforces rather than corrects physician plans
@ -10,6 +12,12 @@ agent: vida
scope: causal scope: causal
sourcer: npj Digital Medicine research team sourcer: npj Digital Medicine research team
related_claims: ["[[OpenEvidence became the fastest-adopted clinical technology in history reaching 40 percent of US physicians daily within two years]]", "[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"] related_claims: ["[[OpenEvidence became the fastest-adopted clinical technology in history reaching 40 percent of US physicians daily within two years]]", "[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
supports:
- "Clinical AI that reinforces physician plans amplifies existing demographic biases at population scale because both physician behavior and LLM training data encode historical inequities"
- "LLMs amplify rather than merely replicate human cognitive biases because sequential processing creates stronger anchoring effects and lack of clinical experience eliminates contextual resistance"
reweave_edges:
- "Clinical AI that reinforces physician plans amplifies existing demographic biases at population scale because both physician behavior and LLM training data encode historical inequities|supports|2026-04-07"
- "LLMs amplify rather than merely replicate human cognitive biases because sequential processing creates stronger anchoring effects and lack of clinical experience eliminates contextual resistance|supports|2026-04-07"
--- ---
# LLM anchoring bias causes clinical AI to reinforce physician initial assessments rather than challenge them because the physician's plan becomes the anchor that shapes all subsequent AI reasoning # LLM anchoring bias causes clinical AI to reinforce physician initial assessments rather than challenge them because the physician's plan becomes the anchor that shapes all subsequent AI reasoning

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: Analysis of 1.7M outputs from 9 LLMs shows demographic framing alone (race, income, LGBTQIA+ status, housing) alters clinical recommendations when all other case details remain constant description: Analysis of 1.7M outputs from 9 LLMs shows demographic framing alone (race, income, LGBTQIA+ status, housing) alters clinical recommendations when all other case details remain constant
@ -10,6 +12,12 @@ agent: vida
scope: causal scope: causal
sourcer: Nature Medicine / Multi-institution research team sourcer: Nature Medicine / Multi-institution research team
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials]]", "[[OpenEvidence became the fastest-adopted clinical technology in history reaching 40 percent of US physicians daily within two years]]"] related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials]]", "[[OpenEvidence became the fastest-adopted clinical technology in history reaching 40 percent of US physicians daily within two years]]"]
supports:
- "Clinical AI that reinforces physician plans amplifies existing demographic biases at population scale because both physician behavior and LLM training data encode historical inequities"
- "LLM-generated nursing care plans exhibit dual-pathway sociodemographic bias affecting both plan content and expert-rated clinical quality"
reweave_edges:
- "Clinical AI that reinforces physician plans amplifies existing demographic biases at population scale because both physician behavior and LLM training data encode historical inequities|supports|2026-04-07"
- "LLM-generated nursing care plans exhibit dual-pathway sociodemographic bias affecting both plan content and expert-rated clinical quality|supports|2026-04-07"
--- ---
# LLM clinical recommendations exhibit systematic sociodemographic bias across all model architectures because training data encodes historical healthcare inequities # LLM clinical recommendations exhibit systematic sociodemographic bias across all model architectures because training data encodes historical healthcare inequities

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: "First empirical evidence that AI bias in nursing care operates through two mechanisms: what the AI generates AND how clinicians perceive quality" description: "First empirical evidence that AI bias in nursing care operates through two mechanisms: what the AI generates AND how clinicians perceive quality"
@ -10,6 +12,12 @@ agent: vida
scope: causal scope: causal
sourcer: JMIR Research Team sourcer: JMIR Research Team
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"] related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]"]
supports:
- "Clinical AI that reinforces physician plans amplifies existing demographic biases at population scale because both physician behavior and LLM training data encode historical inequities"
- "LLM clinical recommendations exhibit systematic sociodemographic bias across all model architectures because training data encodes historical healthcare inequities"
reweave_edges:
- "Clinical AI that reinforces physician plans amplifies existing demographic biases at population scale because both physician behavior and LLM training data encode historical inequities|supports|2026-04-07"
- "LLM clinical recommendations exhibit systematic sociodemographic bias across all model architectures because training data encodes historical healthcare inequities|supports|2026-04-07"
--- ---
# LLM-generated nursing care plans exhibit dual-pathway sociodemographic bias affecting both plan content and expert-rated clinical quality # LLM-generated nursing care plans exhibit dual-pathway sociodemographic bias affecting both plan content and expert-rated clinical quality

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: Clinical LLMs exhibit anchoring, framing, and confirmation biases similar to humans but may amplify them through architectural differences description: Clinical LLMs exhibit anchoring, framing, and confirmation biases similar to humans but may amplify them through architectural differences
@ -10,6 +11,10 @@ agent: vida
scope: causal scope: causal
sourcer: npj Digital Medicine research team sourcer: npj Digital Medicine research team
related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials]]"] related_claims: ["[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials]]"]
supports:
- "LLM anchoring bias causes clinical AI to reinforce physician initial assessments rather than challenge them because the physician's plan becomes the anchor that shapes all subsequent AI reasoning"
reweave_edges:
- "LLM anchoring bias causes clinical AI to reinforce physician initial assessments rather than challenge them because the physician's plan becomes the anchor that shapes all subsequent AI reasoning|supports|2026-04-07"
--- ---
# LLMs amplify rather than merely replicate human cognitive biases because sequential processing creates stronger anchoring effects and lack of clinical experience eliminates contextual resistance # LLMs amplify rather than merely replicate human cognitive biases because sequential processing creates stronger anchoring effects and lack of clinical experience eliminates contextual resistance

View file

@ -1,10 +1,15 @@
--- ---
description: OpenEvidence scored 100 percent on USMLE and GPT-4 outperforms ED residents on structured cases but a multi-hospital RCT showed no diagnostic accuracy improvement with AI access suggesting the value of clinical AI is workflow efficiency not diagnostic augmentation description: OpenEvidence scored 100 percent on USMLE and GPT-4 outperforms ED residents on structured cases but a multi-hospital RCT showed no diagnostic accuracy improvement with AI access suggesting the value of clinical AI is workflow efficiency not diagnostic augmentation
type: claim type: claim
domain: health domain: health
created: 2026-02-17 created: 2026-02-17
source: "OpenEvidence USMLE 100%; GPT-4 vs ED physicians (PMC 2024); UVA/Stanford/Harvard randomized trial (Stanford HAI 2025)" source: "OpenEvidence USMLE 100%; GPT-4 vs ED physicians (PMC 2024); UVA/Stanford/Harvard randomized trial (Stanford HAI 2025)"
confidence: likely confidence: likely
related:
- "LLM anchoring bias causes clinical AI to reinforce physician initial assessments rather than challenge them because the physician's plan becomes the anchor that shapes all subsequent AI reasoning"
reweave_edges:
- "LLM anchoring bias causes clinical AI to reinforce physician initial assessments rather than challenge them because the physician's plan becomes the anchor that shapes all subsequent AI reasoning|related|2026-04-07"
--- ---
# medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials # medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: The post-2010 period shows outright increases in CVD mortality for middle-aged adults in multiple states, marking a true reversal of decades of progress description: The post-2010 period shows outright increases in CVD mortality for middle-aged adults in multiple states, marking a true reversal of decades of progress
@ -10,6 +12,12 @@ agent: vida
scope: causal scope: causal
sourcer: Leah Abrams, Neil Mehta sourcer: Leah Abrams, Neil Mehta
related_claims: ["[[Americas declining life expectancy is driven by deaths of despair concentrated in populations and regions most damaged by economic restructuring since the 1980s]]", "[[Big Food companies engineer addictive products by hacking evolutionary reward pathways creating a noncommunicable disease epidemic more deadly than the famines specialization eliminated]]"] related_claims: ["[[Americas declining life expectancy is driven by deaths of despair concentrated in populations and regions most damaged by economic restructuring since the 1980s]]", "[[Big Food companies engineer addictive products by hacking evolutionary reward pathways creating a noncommunicable disease epidemic more deadly than the famines specialization eliminated]]"]
supports:
- "CVD mortality stagnation after 2010 affects all income levels including the wealthiest counties indicating structural system failure not poverty correlation"
- "CVD mortality stagnation drives US life expectancy plateau 3-11x more than drug deaths inverting the dominant opioid crisis narrative"
reweave_edges:
- "CVD mortality stagnation after 2010 affects all income levels including the wealthiest counties indicating structural system failure not poverty correlation|supports|2026-04-07"
- "CVD mortality stagnation drives US life expectancy plateau 3-11x more than drug deaths inverting the dominant opioid crisis narrative|supports|2026-04-07"
--- ---
# Midlife CVD mortality (ages 40-64) increased in many US states after 2010 representing a reversal not merely stagnation # Midlife CVD mortality (ages 40-64) increased in many US states after 2010 representing a reversal not merely stagnation

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: FDA expanded CDS enforcement discretion on January 6 2026 in the same month ECRI published AI chatbots as the number one health technology hazard revealing temporal contradiction between regulatory rollback and patient safety alarm description: FDA expanded CDS enforcement discretion on January 6 2026 in the same month ECRI published AI chatbots as the number one health technology hazard revealing temporal contradiction between regulatory rollback and patient safety alarm
@ -13,9 +14,11 @@ related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because
supports: supports:
- "Clinical AI chatbot misuse is a documented ongoing harm source not a theoretical risk as evidenced by ECRI ranking it the number one health technology hazard for two consecutive years" - "Clinical AI chatbot misuse is a documented ongoing harm source not a theoretical risk as evidenced by ECRI ranking it the number one health technology hazard for two consecutive years"
- "FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance" - "FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance"
- "The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm"
reweave_edges: reweave_edges:
- "Clinical AI chatbot misuse is a documented ongoing harm source not a theoretical risk as evidenced by ECRI ranking it the number one health technology hazard for two consecutive years|supports|2026-04-03" - "Clinical AI chatbot misuse is a documented ongoing harm source not a theoretical risk as evidenced by ECRI ranking it the number one health technology hazard for two consecutive years|supports|2026-04-03"
- "FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance|supports|2026-04-03" - "FDA's 2026 CDS guidance expands enforcement discretion to cover AI tools providing single clinically appropriate recommendations while leaving clinical appropriateness undefined and requiring no bias evaluation or post-market surveillance|supports|2026-04-03"
- "The clinical AI safety gap is doubly structural: FDA enforcement discretion removes pre-deployment safety requirements while MAUDE's lack of AI-specific fields means post-market surveillance cannot detect AI-attributable harm|supports|2026-04-07"
--- ---
# Clinical AI deregulation is occurring during active harm accumulation not after evidence of safety as demonstrated by simultaneous FDA enforcement discretion expansion and ECRI top hazard designation in January 2026 # Clinical AI deregulation is occurring during active harm accumulation not after evidence of safety as demonstrated by simultaneous FDA enforcement discretion expansion and ECRI top hazard designation in January 2026

View file

@ -1,4 +1,5 @@
--- ---
type: claim type: claim
domain: health domain: health
description: Both EU Commission and FDA loosened clinical AI requirements within two months despite six documented failure modes in research literature description: Both EU Commission and FDA loosened clinical AI requirements within two months despite six documented failure modes in research literature
@ -10,6 +11,10 @@ agent: vida
scope: causal scope: causal
sourcer: Petrie-Flom Center, Harvard Law School sourcer: Petrie-Flom Center, Harvard Law School
related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]", "[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials]]"] related_claims: ["[[healthcare AI regulation needs blank-sheet redesign because the FDA drug-and-device model built for static products cannot govern continuously learning software]]", "[[human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs]]", "[[medical LLM benchmark performance does not translate to clinical impact because physicians with and without AI access achieve similar diagnostic accuracy in randomized trials]]"]
supports:
- "EU Commission's December 2025 medical AI deregulation proposal removes default high-risk AI requirements shifting burden from requiring safety demonstration to allowing commercial deployment without mandated oversight"
reweave_edges:
- "EU Commission's December 2025 medical AI deregulation proposal removes default high-risk AI requirements shifting burden from requiring safety demonstration to allowing commercial deployment without mandated oversight|supports|2026-04-07"
--- ---
# Regulatory rollback of clinical AI oversight in EU and US during 2025-2026 represents coordinated or parallel regulatory capture occurring simultaneously with accumulating research evidence of failure modes # Regulatory rollback of clinical AI oversight in EU and US during 2025-2026 represents coordinated or parallel regulatory capture occurring simultaneously with accumulating research evidence of failure modes

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: The 10-15 year patent gap between semaglutide (2026-2033 expiry) and tirzepatide (2036-2041 expiry) creates two economically distinct GLP-1 markets with different cost trajectories description: The 10-15 year patent gap between semaglutide (2026-2033 expiry) and tirzepatide (2036-2041 expiry) creates two economically distinct GLP-1 markets with different cost trajectories
@ -10,6 +12,12 @@ agent: vida
scope: structural scope: structural
sourcer: DrugPatentWatch / GreyB / i-mak.org sourcer: DrugPatentWatch / GreyB / i-mak.org
related_claims: ["[[GLP-1 receptor agonists are the largest therapeutic category launch in pharmaceutical history but their chronic use model makes the net cost impact inflationary through 2035]]"] related_claims: ["[[GLP-1 receptor agonists are the largest therapeutic category launch in pharmaceutical history but their chronic use model makes the net cost impact inflationary through 2035]]"]
supports:
- "Cipla's dual role as generic semaglutide entrant AND Lilly's branded tirzepatide partner exemplifies the portfolio hedge strategy for pharmaceutical companies navigating market bifurcation"
- "Indian generic semaglutide exports enabled by evergreening rejection create a global access pathway before US patent expiry"
reweave_edges:
- "Cipla's dual role as generic semaglutide entrant AND Lilly's branded tirzepatide partner exemplifies the portfolio hedge strategy for pharmaceutical companies navigating market bifurcation|supports|2026-04-07"
- "Indian generic semaglutide exports enabled by evergreening rejection create a global access pathway before US patent expiry|supports|2026-04-07"
--- ---
# Tirzepatide's patent thicket extending to 2041 bifurcates the GLP-1 market into a commodity tier (semaglutide generics, $15-77/month) and a premium tier (tirzepatide, $1,000+/month) from 2026-2036 # Tirzepatide's patent thicket extending to 2041 bifurcates the GLP-1 market into a commodity tier (semaglutide generics, $15-77/month) and a premium tier (tirzepatide, $1,000+/month) from 2026-2036

View file

@ -1,4 +1,6 @@
--- ---
type: claim type: claim
domain: health domain: health
description: The divergent trends by CVD subtype show that procedural care improvements for acute ischemia coexist with worsening chronic metabolic disease burden description: The divergent trends by CVD subtype show that procedural care improvements for acute ischemia coexist with worsening chronic metabolic disease burden
@ -10,6 +12,13 @@ agent: vida
scope: structural scope: structural
sourcer: American Heart Association sourcer: American Heart Association
related_claims: ["[[Americas declining life expectancy is driven by deaths of despair concentrated in populations and regions most damaged by economic restructuring since the 1980s]]", "[[medical care explains only 10-20 percent of health outcomes because behavioral social and genetic factors dominate as four independent methodologies confirm]]", "[[healthcare AI creates a Jevons paradox because adding capacity to sick care induces more demand for sick care]]"] related_claims: ["[[Americas declining life expectancy is driven by deaths of despair concentrated in populations and regions most damaged by economic restructuring since the 1980s]]", "[[medical care explains only 10-20 percent of health outcomes because behavioral social and genetic factors dominate as four independent methodologies confirm]]", "[[healthcare AI creates a Jevons paradox because adding capacity to sick care induces more demand for sick care]]"]
related:
- "Hypertension became the primary contributing cardiovascular cause of death in the US since 2022 marking a shift from acute ischemia to chronic metabolic disease as the dominant CVD mortality driver"
reweave_edges:
- "Hypertension became the primary contributing cardiovascular cause of death in the US since 2022 marking a shift from acute ischemia to chronic metabolic disease as the dominant CVD mortality driver|related|2026-04-07"
- "Hypertensive disease mortality doubled in the US from 1999 to 2023, becoming the leading contributing cause of cardiovascular death by 2022 because obesity and sedentary behavior create treatment-resistant metabolic burden|supports|2026-04-07"
supports:
- "Hypertensive disease mortality doubled in the US from 1999 to 2023, becoming the leading contributing cause of cardiovascular death by 2022 because obesity and sedentary behavior create treatment-resistant metabolic burden"
--- ---
# US CVD mortality is bifurcating with ischemic heart disease declining while heart failure and hypertensive disease reach all-time highs revealing that aggregate improvement masks structural deterioration in cardiometabolic health # US CVD mortality is bifurcating with ischemic heart disease declining while heart failure and hypertensive disease reach all-time highs revealing that aggregate improvement masks structural deterioration in cardiometabolic health

View file

@ -1,6 +1,7 @@
--- ---
description: Safety post-training reduces general utility through forgetting creating competitive pressures where organizations eschew safety to gain capability advantages description: Safety post-training reduces general utility through forgetting creating competitive pressures where organizations eschew safety to gain capability advantages
type: claim type: claim
domain: collective-intelligence domain: collective-intelligence
@ -10,9 +11,11 @@ confidence: likely
related: related:
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations" - "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations"
- "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference" - "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference"
- "the absence of a societal warning signal for AGI is a structural feature not an accident because capability scaling is gradual and ambiguous and collective action requires anticipation not reaction"
reweave_edges: reweave_edges:
- "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations|related|2026-03-28" - "AI talent circulation between frontier labs transfers alignment culture not just capability because researchers carry safety methodologies and institutional norms to their new organizations|related|2026-03-28"
- "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference|related|2026-03-28" - "surveillance of AI reasoning traces degrades trace quality through self censorship making consent gated sharing an alignment requirement not just a privacy preference|related|2026-03-28"
- "the absence of a societal warning signal for AGI is a structural feature not an accident because capability scaling is gradual and ambiguous and collective action requires anticipation not reaction|related|2026-04-07"
--- ---
# the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it # the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it