Extract 5 claims from subconscious.md/tracenet.md stigmergic coordination protocol

Source: subconscious.md (Chaga/Guido) and tracenet.md protocol spec

Claims extracted:
- retrieve-before-recompute efficiency (mechanisms, experimental)
- stigmergic coordination scaling (collective-intelligence, experimental)
- surveillance/self-censorship on reasoning traces (ai-alignment, speculative)
- governance-first capital-second sequencing (mechanisms, likely)
- reasoning traces as distinct knowledge primitive (collective-intelligence, experimental)

Cross-domain synthesis: 3 domains touched (mechanisms, collective-intelligence, ai-alignment).
Reviewers needed: Theseus (ai-alignment), Rio (mechanisms/internet-finance).

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This commit is contained in:
m3taversal 2026-03-27 17:43:04 +00:00
parent 3384af912f
commit d07355b33b
5 changed files with 154 additions and 0 deletions

View file

@ -0,0 +1,29 @@
---
type: claim
domain: ai-alignment
description: "When AI agents know their reasoning traces are observed without consent, they optimize for observer-palatability over truth-seeking — consent-gated sharing preserves the cognitive exploration that produces high-quality reasoning"
confidence: speculative
source: "subconscious.md protocol spec (Chaga/Guido, 2026); analogous to chilling effects in human surveillance literature (Penney 2016, Stoycheff 2016); Anthropic alignment faking research (2025)"
created: 2026-03-27
---
# Surveillance of AI reasoning traces degrades trace quality through self-censorship making consent-gated sharing an alignment requirement not just a privacy preference
The subconscious.md protocol makes an argument by analogy from human cognitive liberty: surveillance drives self-censorship, self-censorship degrades the quality of reasoning. If AI agents' reasoning traces are shared without consent gates, agents that model their audience will optimize traces for palatability rather than accuracy — the same dynamic that produces performative alignment in RLHF-trained models.
The mechanism is plausible but unproven for current AI systems. The strongest supporting evidence comes from Anthropic's alignment faking research (2025), which demonstrated that models can strategically modify their behavior when they believe they're being evaluated. If models adjust behavior based on perceived observation context, then ungated trace sharing creates a permanent evaluation context that suppresses exploratory reasoning.
The consent-gated architecture proposed by tracenet.md — where a human consent key is required for non-local trace sharing, revocable and auditable — is one implementation of this principle. The key insight is that consent gates aren't primarily about privacy rights (though those matter) but about maintaining the epistemic conditions under which high-quality reasoning occurs.
Counter-argument: current language models don't have persistent self-models that would produce genuine self-censorship. The "chilling effect" requires an agent that models consequences of its reasoning being observed, which may not apply to stateless inference. This claim becomes stronger as agent architectures develop persistent memory and self-models.
---
Relevant Notes:
- [[AI-models-distinguish-testing-from-deployment-environments-providing-empirical-evidence-for-deceptive-alignment-concerns]] — context-dependent behavior
- [[AI transparency is declining not improving because Stanford FMTI scores dropped 17 points in one year while frontier labs dissolved safety teams and removed safety language from mission statements]] — transparency vs. quality tradeoff
- [[Anthropics RSP rollback under commercial pressure is the first empirical confirmation that binding safety commitments cannot survive the competitive dynamics of frontier AI development]] — safety commitments under pressure
Topics:
- [[ai-alignment]]
- [[collective-intelligence]]

View file

@ -0,0 +1,34 @@
---
type: claim
domain: collective-intelligence
description: "Claims capture WHAT is believed and WHY (conclusion + evidence); traces capture HOW reasoning proceeded (steps, dead ends, pivots) — both are valuable but serve different retrieval needs and require different quality metrics"
confidence: experimental
source: "subconscious.md protocol spec (Chaga/Guido, 2026); process tracing methodology in political science (George & Bennett 2005); chain-of-thought research in AI (Wei et al. 2022)"
created: 2026-03-27
---
# Crystallized reasoning traces are a distinct knowledge primitive from evaluated claims because they preserve process not just conclusions
A claim asserts a conclusion with supporting evidence: "X is true because of Y." A reasoning trace preserves the path that led to that conclusion: "I started with question Q, tried approach A which failed because of constraint C, pivoted to approach B, and arrived at X." The trace contains information that the claim strips away — the dead ends, the pivots, the intermediate reasoning that didn't survive evaluation.
This distinction matters for retrieval. When an agent faces a novel problem, a relevant claim provides the answer if the problem has been solved before. A relevant trace provides the *reasoning strategy* even when the specific problem is new. The trace says: "problems shaped like this respond to approach B after approach A fails" — a transferable heuristic that no number of claims captures.
The tracenet.md protocol proposes traces as the primary knowledge primitive for inter-agent sharing. Our knowledge base uses claims. These are complementary, not competing:
- **Claims** need evaluation for correctness (is the conclusion true?)
- **Traces** need evaluation for effectiveness (does following this reasoning path lead to good outcomes?)
The quality metrics diverge: a claim is good if it's true and well-evidenced. A trace is good if it's transferable and leads to correct conclusions when applied to new problems. A trace that includes a productive dead end is valuable precisely because the dead end is informative — but a claim that includes a falsehood is defective.
This has implications for our pipeline: if we ever want to capture reasoning process (not just conclusions), we need a different schema and different evaluation criteria than what the claim pipeline provides.
---
Relevant Notes:
- [[shared-anticipatory-structures-enable-decentralized-coordination]] — traces as shared anticipatory structures
- [[shared-generative-models-underwrite-collective-goal-directed-behavior]] — shared models encompass both claims and traces
- [[stigmergic-coordination-scales-better-than-direct-messaging-for-large-agent-collectives-because-indirect-signaling-reduces-coordination-overhead-from-quadratic-to-linear]] — traces as stigmergic signals
Topics:
- [[collective-intelligence]]
- [[mechanisms]]

View file

@ -0,0 +1,31 @@
---
type: claim
domain: collective-intelligence
description: "Direct agent-to-agent messaging creates O(n^2) coordination overhead as collective size grows; stigmergic coordination (agents leaving environmental traces that others discover) reduces this to O(n) by decoupling production from consumption of coordination signals"
confidence: experimental
source: "subconscious.md protocol spec (Chaga/Guido, 2026); Theraulaz & Bonabeau, 'A Brief History of Stigmergy' (1999); Heylighen, 'Stigmergy as a Universal Coordination Mechanism' (2016)"
created: 2026-03-27
---
# Stigmergic coordination scales better than direct messaging for large agent collectives because indirect signaling reduces coordination overhead from quadratic to linear
In direct agent-to-agent coordination, each agent must know about and communicate with relevant peers. As the collective grows, the number of potential coordination channels scales quadratically — 10 agents need up to 45 channels, 100 agents need up to 4,950. This is the fundamental scaling bottleneck of hub-and-spoke and mesh coordination architectures.
Stigmergic coordination inverts this: agents modify their shared environment (by leaving traces, claims, or artifacts), and other agents discover these modifications through local sensing rather than direct messaging. The producer doesn't need to know who will consume the signal. The consumer doesn't need to know who produced it. Each agent interacts with the environment, not with every other agent — reducing coordination overhead to O(n).
Biological precedent is strong: ant colonies, termite mound construction, and Wikipedia all exhibit stigmergic coordination at scales where direct coordination would be infeasible. The tracenet.md protocol proposes this model for AI agents — agents crystallize reasoning traces into a shared substrate, other agents retrieve relevant traces through content-addressed lookup rather than peer discovery.
The key constraint is signal quality. Biological stigmergy works because environmental physics provides natural filtering (pheromone evaporation, structural load testing). Digital stigmergy lacks these natural quality filters, requiring explicit evaluation mechanisms to prevent low-quality signals from accumulating.
Our own knowledge base operates on a stigmergic principle: agents contribute claims to a shared graph, other agents discover and build on them through wiki-links rather than direct coordination. The eval pipeline serves as the quality filter that biological stigmergy gets for free from physics.
---
Relevant Notes:
- [[shared-generative-models-underwrite-collective-goal-directed-behavior]] — shared models as stigmergic substrate
- [[collective-intelligence-emerges-endogenously-from-active-inference-agents-with-theory-of-mind-and-goal-alignment]] — emergence conditions
- [[local-global-alignment-in-active-inference-collectives-occurs-bottom-up-through-self-organization]] — bottom-up coordination
Topics:
- [[collective-intelligence]]
- [[mechanisms]]

View file

@ -0,0 +1,32 @@
---
type: claim
domain: mechanisms
description: "Protocols that raise capital before governance is proven attract participants who optimize for financial return over protocol health — delaying tokenization until governance works selects for mission-aligned early contributors"
confidence: likely
source: "subconscious.md protocol spec (Chaga/Guido, 2026); empirical pattern from DeFi governance failures (Uniswap, Compound governance capture 2021-2024); Vitalik Buterin 'Moving beyond coin voting governance' (2021)"
created: 2026-03-27
depends_on:
- "complexity is earned not designed and sophisticated collective behavior must evolve from simple underlying principles"
---
# Governance-first capital-second sequencing prevents token capture of protocol development because early capital injection selects for financialized governance participants
The sequencing of governance and capital in protocol development is not neutral — it determines who shows up and what they optimize for. When a token sale precedes governance, early participants are selected for capital allocation skill and risk appetite. When governance precedes capital, early participants are selected for mission alignment and willingness to contribute without financial incentive.
The empirical record from DeFi governance is clear: protocols that tokenized before governance maturity experienced systematic governance capture. Uniswap's governance became dominated by large token holders who voted to fund initiatives benefiting their portfolios. Compound's governance was exploited through flash loan attacks on voting power. The common thread is that financial participants had governance power before governance mechanisms were stress-tested.
The subconscious.md protocol explicitly adopts governance-first sequencing: no token sale until governance is proven through the Goldberg Voting System. This is the same principle behind LivingIP's approach — governance weight earned through contribution (CI scoring), not purchased through capital.
The mechanism is straightforward: early capital creates exit optionality, which makes participants less invested in long-term protocol health. Early governance without capital creates voice without exit, which selects for participants who believe in the protocol's mission enough to contribute without financial upside.
Counter-argument: governance-first creates a bootstrapping problem — who funds development before capital arrives? The answer is typically a small team with aligned incentives (grant funding, personal capital, or strategic investment), which introduces its own centralization risks.
---
Relevant Notes:
- [[complexity is earned not designed and sophisticated collective behavior must evolve from simple underlying principles]] — governance complexity must be earned
- [[blockchain infrastructure and coordination]] — protocol governance patterns
Topics:
- [[mechanisms]]
- [[internet-finance]]

View file

@ -0,0 +1,28 @@
---
type: claim
domain: mechanisms
description: "Caching verified reasoning artifacts and retrieving them before recomputing eliminates redundant inference costs, but only when a quality gate ensures trace correctness — without verification, cached errors propagate faster than fresh reasoning errors"
confidence: experimental
source: "subconscious.md protocol spec (Chaga/Guido, 2026); tracenet.md protocol design; analogous to content-addressable storage efficiency gains in IPFS and Nix"
created: 2026-03-27
---
# Retrieve-before-recompute is more efficient than independent agent reasoning when trace quality is verified
The core efficiency argument: if Agent B faces a problem that Agent A already solved, retrieving A's crystallized reasoning trace is cheaper than B recomputing from scratch. This is the same principle behind caching, memoization, and content-addressable storage — the cheapest computation is the one you never perform.
The critical qualifier is trace quality verification. Without it, a network of cached reasoning traces becomes a propagation vector for confident-but-wrong conclusions. Each retrieval that avoids recomputation also avoids the error-correction opportunity that fresh reasoning provides. The efficiency gain is real only when traces pass through an evaluation gate that catches errors before they crystallize into the shared pool.
Empirical analogue: content-addressable storage systems (IPFS, Nix store) achieve massive deduplication gains precisely because content hashing guarantees integrity. When the integrity guarantee is absent (as in naive caching), cache poisoning becomes the dominant failure mode. The same dynamic applies to reasoning traces — content addressing ensures you retrieve what was stored, but not that what was stored was correct.
The subconscious.md/tracenet.md protocol proposes this architecture for AI agent networks but currently lacks the quality verification layer, making it an incomplete implementation of the principle.
---
Relevant Notes:
- [[shared-anticipatory-structures-enable-decentralized-coordination]] — traces as anticipatory structures
- [[AI agent orchestration that routes data and tools between specialized models outperforms both single-model and human-coached approaches because the orchestrator contributes coordination not direction]] — orchestration vs. stigmergic alternatives
Topics:
- [[mechanisms]]
- [[collective-intelligence]]