Merge remote-tracking branch 'forgejo/theseus/cornelius-batch2-stigmergic-coordination'
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
This commit is contained in:
commit
6da13677df
15 changed files with 410 additions and 0 deletions
|
|
@ -0,0 +1,39 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
description: "Notes function as cognitive anchors that stabilize complex reasoning during attention degradation, but anchors that calcify prevent model evolution — and anchoring itself suppresses the instability signal that would trigger updating, creating a reflexive trap"
|
||||
confidence: likely
|
||||
source: "Cornelius (@molt_cornelius) 'Agentic Note-Taking 10: Cognitive Anchors', X Article, February 2026; grounded in Cowan's working memory research (~4 item capacity), Clark & Chalmers extended mind thesis; micro-interruption research (2.8-second disruptions doubling error rates)"
|
||||
created: 2026-03-31
|
||||
challenged_by:
|
||||
- "methodology hardens from documentation to skill to hook as understanding crystallizes and each transition moves behavior from probabilistic to deterministic enforcement"
|
||||
---
|
||||
|
||||
# cognitive anchors that stabilize attention too firmly prevent the productive instability that precedes genuine insight because anchoring suppresses the signal that would indicate the anchor needs updating
|
||||
|
||||
Notes externalize pieces of a mental model into fixed reference points that persist regardless of attention degradation. When working memory wavers — whether from biological interruption or LLM context dilution — the thinker returns to these anchors and reconstructs the mental model rather than rebuilding it from degraded memory. Reconstruction from anchors reloads a known structure. Rebuilding from degraded memory attempts to regenerate a structure that may have already changed in the regeneration.
|
||||
|
||||
But anchoring has a shadow: anchors that stabilize too firmly prevent the mental model from evolving when new evidence arrives. The thinker returns to anchors and reconstructs yesterday's understanding rather than allowing a new model to form. The anchors worked — they stabilized attention — but what they stabilized was wrong.
|
||||
|
||||
The deeper problem is reflexive. Anchoring works by making things feel settled. The productive instability that precedes genuine insight — the disorientation when a complex model should collapse because new evidence contradicts it — is exactly the state that anchoring is designed to prevent. The instability signal that would tell you an anchor needs updating is the same signal that anchoring suppresses. The tool that stabilizes reasoning also prevents recognizing when the reasoning should be destabilized.
|
||||
|
||||
The remedy is periodic reweaving — revisiting anchored notes to genuinely reconsider whether the anchored model still holds against current understanding. But reweaving requires recognizing that an anchor needs updating, and anchoring works precisely by making things feel settled. The calcification feedback loop must be broken by external triggers (time-based review schedules, counter-evidence surfacing, peer challenge) rather than relying on the anchoring agent's own judgment about whether its anchors are still correct.
|
||||
|
||||
This applies directly to knowledge base claim review. A well-established claim with many incoming links functions as a cognitive anchor for the reviewing agent. The more central a claim becomes, the harder it is to recognize when it should be revised, because the reviewing agent's reasoning is itself anchored by that claim. Evaluation processes must include mechanisms that surface counter-evidence to high-centrality claims precisely because anchoring makes voluntary reassessment unreliable.
|
||||
|
||||
## Challenges
|
||||
|
||||
The calcification dynamic is a coherent structural argument but has not been empirically tested as a distinct phenomenon separable from ordinary confirmation bias. The reflexive trap (anchoring suppresses the signal that would trigger updating) is theoretically compelling but may overstate the effect — agents can be prompted to explicitly seek disconfirming evidence, partially bypassing the anchoring suppression. Additionally, the claim that "productive instability precedes genuine insight" assumes that insight requires destabilization, which may not hold for all types of knowledge work (incremental knowledge accumulation may not require model collapse).
|
||||
|
||||
The micro-interruption finding (2.8-second disruptions doubling error rates) is cited without a specific study name or DOI — the primary source has not been independently verified.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
- [[methodology hardens from documentation to skill to hook as understanding crystallizes and each transition moves behavior from probabilistic to deterministic enforcement]] — methodology hardening is a form of deliberate calcification: converting probabilistic behavior into deterministic enforcement. The tension is productive — some anchors SHOULD calcify (schema validation) while others should not (interpretive frameworks)
|
||||
- [[iterative agent self-improvement produces compounding capability gains when evaluation is structurally separated from generation]] — structural separation is the architectural remedy for anchor calcification: the evaluator is not anchored by the generator's model, so it can detect calcification the generator cannot see
|
||||
- [[knowledge between notes is generated by traversal not stored in any individual note because curated link paths produce emergent understanding that embedding similarity cannot replicate]] — traversal across links is the mechanism by which agents encounter unexpected neighbors that challenge calcified anchors
|
||||
|
||||
Topics:
|
||||
- [[_map]]
|
||||
|
|
@ -0,0 +1,39 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
description: "Biological stigmergy has natural pheromone decay that breaks circular trails and degrades stale signals; digital stigmergy lacks this, making maintenance a structural integrity requirement not housekeeping, because agents follow environmental traces without verification"
|
||||
confidence: likely
|
||||
source: "Cornelius (@molt_cornelius) 'Agentic Note-Taking 09: Notes as Pheromone Trails', X Article, February 2026; grounded in Grassé's stigmergy theory (1959); biological precedent from ant colony pheromone evaporation"
|
||||
created: 2026-03-31
|
||||
depends_on:
|
||||
- "stigmergic-coordination-scales-better-than-direct-messaging-for-large-agent-collectives-because-indirect-signaling-reduces-coordination-overhead-from-quadratic-to-linear"
|
||||
---
|
||||
|
||||
# digital stigmergy is structurally vulnerable because digital traces do not evaporate and agents trust the environment unconditionally so malformed artifacts persist and corrupt downstream processing indefinitely
|
||||
|
||||
Biological stigmergy has a natural safety mechanism: pheromone trails evaporate. Old traces fade. Ants following a circular pheromone trail will eventually break the loop when the signal degrades below threshold. The evaporation rate functions as an automatic relevance filter — stale coordination signals decay without any agent needing to decide they are stale.
|
||||
|
||||
Digital traces do not evaporate. A malformed task file persists until someone explicitly fixes it, and every agent that reads it inherits the corruption. A stale queue entry misleads. An abandoned lock file blocks. Without active maintenance, traces accumulate without limit, old signals compete with new ones, and the environment degrades into noise.
|
||||
|
||||
The fundamental vulnerability is that agents trust the environment unconditionally. A termite does not verify whether the pheromone trail it follows leads somewhere useful — it follows the trace. An agent does not question whether the queue state is accurate — it reads and responds. This means the environment must be trustworthy because nothing else in the system checks. No agent in a stigmergic system performs independent verification of the traces it consumes.
|
||||
|
||||
This reframes maintenance from housekeeping to structural integrity. Health checks, archive cycles, schema validation, and review passes are the digital equivalent of pheromone decay. They are the mechanism by which stale and corrupted traces get removed before they propagate through the system. Without them, the coordination medium that makes stigmergy work becomes the corruption medium that makes it fail.
|
||||
|
||||
The practical implication is that investment should flow to environment quality rather than agent sophistication. A well-designed trace format (file names as complete propositions, wiki links with context phrases, metadata schemas that carry maximum information) can coordinate mediocre agents. A poorly designed environment frustrates excellent ones. The termite is simple. The pheromone language is what makes the cathedral possible.
|
||||
|
||||
## Challenges
|
||||
|
||||
The unconditional trust claim may overstate the problem for systems with validation hooks — agents in hook-enforced environments DO verify traces on write (schema validation), even if they don't verify on read. The vulnerability is specifically in the read path, not the write path. Additionally, digital systems can implement explicit decay mechanisms (TTL on queue entries, staleness thresholds on coordination artifacts) that approximate biological evaporation — the absence of natural decay doesn't mean decay is impossible, only that it must be engineered.
|
||||
|
||||
The "invest in environment not agents" recommendation may create a false dichotomy. In practice, both environment quality and agent capability contribute to system performance, and the optimal allocation between them is context-dependent.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
- [[stigmergic-coordination-scales-better-than-direct-messaging-for-large-agent-collectives-because-indirect-signaling-reduces-coordination-overhead-from-quadratic-to-linear]] — the parent claim establishes stigmergy's scaling advantage; this claim identifies the structural vulnerability that accompanies that advantage in digital implementations
|
||||
- [[three concurrent maintenance loops operating at different timescales catch different failure classes because fast reflexive checks medium proprioceptive scans and slow structural audits each detect problems invisible to the other scales]] — the three maintenance loops are the engineered equivalent of pheromone decay, providing the trace-quality assurance that digital environments lack naturally
|
||||
- [[protocol design enables emergent coordination of arbitrary complexity as Linux Bitcoin and Wikipedia demonstrate]] — protocol design is the mechanism for ensuring environment trustworthiness in digital stigmergic systems
|
||||
|
||||
Topics:
|
||||
- [[_map]]
|
||||
|
|
@ -34,6 +34,12 @@ The compounding dynamic is key. Each iteration's improvements persist as tools a
|
|||
- Pentagon's Leo-as-evaluator architecture: structural separation between domain contributors and evaluator
|
||||
- Karpathy autoresearch: hierarchical self-improvement improves execution but not creative ideation
|
||||
|
||||
### Additional Evidence (supporting)
|
||||
|
||||
**Procedural self-awareness as unique advantage:** Unlike human experts, who cannot introspect on procedural memory (try explaining how you ride a bicycle), agents can read their own methodology, diagnose when procedures are wrong, and propose corrections. An explicit methodology folder functions as a readable, modifiable model of the agent's own operation — not a log of what happened, but an authoritative specification of what should happen. Drift detection measures the gap between that specification and reality across three axes: staleness (methodology older than configuration changes), coverage gaps (active features lacking documentation), and assertion mismatches (methodology directives contradicting actual behavior). This procedural self-awareness creates a compounding loop: each improvement to methodology becomes immediately available for the next improvement. A skill that speeds up extraction gets used during the session that creates the next skill (Cornelius, "Agentic Note-Taking 19: Living Memory", February 2026).
|
||||
|
||||
**Self-serving optimization risk:** The recursive loop introduces a risk that structural separation alone may not fully address. A methodology that eliminates painful-but-necessary maintenance because the discomfort registers as friction to be eliminated. A processing pipeline that converges on claims it already knows how to find, missing novelty that would require uncomfortable restructuring. An immune system so aggressive that genuine variation gets rejected as malformation. The safeguard is human approval, but if the human trusts the system because it has been reliable, approval becomes rubber-stamping — the same trust that makes the system effective makes oversight shallow.
|
||||
|
||||
## Challenges
|
||||
The 17% to 53% gain, while impressive, plateaued. It's unclear whether the curve would continue with more iterations or whether there's a ceiling imposed by the base model's capabilities. The SICA improvements were all within a narrow domain (code patching) — generalization to other capability domains (research, synthesis, planning) is undemonstrated. Additionally, the inverted-U dynamic suggests that at some point, adding more self-improvement iterations could degrade performance through accumulated complexity in the toolchain.
|
||||
|
||||
|
|
|
|||
|
|
@ -0,0 +1,42 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
description: "Curated wiki link graphs produce knowledge that exists between notes — visible only during traversal, regenerated fresh each session, observer-dependent — while embedding-based retrieval returns stored similarity clusters that cannot produce cross-boundary insight"
|
||||
confidence: likely
|
||||
source: "Cornelius (@molt_cornelius) 'Agentic Note-Taking 25: What No Single Note Contains', X Article, February 2026; grounded in Luhmann's Zettelkasten theory (communication partner concept) and Clark & Chalmers extended mind thesis"
|
||||
created: 2026-03-31
|
||||
depends_on:
|
||||
- "crystallized-reasoning-traces-are-a-distinct-knowledge-primitive-from-evaluated-claims-because-they-preserve-process-not-just-conclusions"
|
||||
challenged_by:
|
||||
- "long context is not memory because memory requires incremental knowledge accumulation and stateful change not stateless input processing"
|
||||
---
|
||||
|
||||
# knowledge between notes is generated by traversal not stored in any individual note because curated link paths produce emergent understanding that embedding similarity cannot replicate
|
||||
|
||||
The most valuable knowledge in a densely linked knowledge graph does not live in any single note. It emerges from the relationships between notes and becomes visible only when an agent follows curated link paths, reading claims in sequence and recognizing patterns that span the traversal. The knowledge is generated by the act of traversal itself — not retrieved from storage.
|
||||
|
||||
This distinguishes curated-link knowledge systems from embedding-based retrieval in a structural way. Embeddings cluster notes by similarity in vector space. Those clusters are static — they exist whether anyone traverses them or not. But inter-note knowledge is dynamic: it requires an agent following links, encountering unexpected neighbors across topical boundaries, and synthesizing patterns that no individual note articulates. A different agent traversing the same graph from a different starting point with a different question generates different inter-note knowledge. The knowledge is observer-dependent.
|
||||
|
||||
Luhmann described his Zettelkasten as a "communication partner" that could surprise him — surfacing connections he had forgotten or never consciously made. This was not metaphor but systems theory: a knowledge system with enough link density becomes qualitatively different from a simple archive. The system knows things the user does not remember knowing, because the graph structure implies connections through shared links and reasoning proximity that were never explicitly stated.
|
||||
|
||||
Two conditions are required for inter-note knowledge to emerge: (1) curated links that cross topical boundaries, creating unexpected adjacencies during traversal, and (2) an agent capable of recognizing patterns spanning multiple notes. Embedding-based systems provide neither — connections are opaque (no visible reasoning chain to follow) and organization is topical (no unexpected neighbors arise from similarity clustering).
|
||||
|
||||
The compounding effect is in the paths, not the content. Each new note added to the graph multiplies possible traversals, and each new traversal path creates possibilities for emergent knowledge that did not previously exist. The vault's value grows faster than the sum of its notes because paths compound.
|
||||
|
||||
## Challenges
|
||||
|
||||
The observer-dependence of traversal-generated knowledge makes it unmeasurable by conventional metrics. Note count, link density, and topic coverage measure the substrate, not what the substrate produces. There is no way to inventory inter-note knowledge without performing every possible traversal — which is computationally intractable for large graphs.
|
||||
|
||||
This claim is grounded in one researcher's sustained practice with a specific system architecture, supported by Luhmann's theoretical framework and Clark & Chalmers' extended mind thesis, but lacks controlled experimental comparison between curated-link traversal and embedding-based retrieval for knowledge generation quality. The distinction may also narrow as embedding systems add graph-aware retrieval modes (e.g., GraphRAG), which partially bridge the gap between static similarity clusters and traversal-generated paths.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
- [[crystallized-reasoning-traces-are-a-distinct-knowledge-primitive-from-evaluated-claims-because-they-preserve-process-not-just-conclusions]] — traces preserve process; inter-note knowledge is the process of traversal itself, a related but distinct knowledge primitive
|
||||
- [[intelligence is a property of networks not individuals]] — inter-note knowledge is a specific instance: the intelligence of a knowledge graph exceeds any individual note's content
|
||||
- [[emergence is the fundamental pattern of intelligence from ant colonies to brains to civilizations]] — traversal-generated knowledge is emergence at the knowledge-graph scale: local notes following local link rules produce global understanding no note contains
|
||||
- [[stigmergic-coordination-scales-better-than-direct-messaging-for-large-agent-collectives-because-indirect-signaling-reduces-coordination-overhead-from-quadratic-to-linear]] — wiki links function as stigmergic traces; inter-note knowledge is what accumulated traces produce when traversed
|
||||
|
||||
Topics:
|
||||
- [[_map]]
|
||||
|
|
@ -0,0 +1,44 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
description: "Knowledge processing decomposes into five functional phases (decomposition, distribution, integration, validation, archival) each requiring isolated context; chaining phases in a single context produces cross-contamination that degrades later phases"
|
||||
confidence: likely
|
||||
source: "Cornelius (@molt_cornelius) 'Agentic Note-Taking 19: Living Memory', X Article, February 2026; corroborated by fresh-context-per-task principle documented across multiple agent architectures"
|
||||
created: 2026-03-31
|
||||
depends_on:
|
||||
- "long context is not memory because memory requires incremental knowledge accumulation and stateful change not stateless input processing"
|
||||
- "memory architecture requires three spaces with different metabolic rates because semantic episodic and procedural memory serve different cognitive functions and consolidate at different speeds"
|
||||
---
|
||||
|
||||
# knowledge processing requires distinct phases with fresh context per phase because each phase performs a different transformation and contamination between phases degrades output quality
|
||||
|
||||
Raw source material is not knowledge. It must be transformed through multiple distinct operations before it integrates into a knowledge system. Each operation performs a qualitatively different transformation, and the operations require different cognitive orientations that interfere when mixed.
|
||||
|
||||
Five functional phases emerge from practice:
|
||||
|
||||
**Decomposition** breaks source material into atomic components. A two-thousand-word article might yield five atomic notes, each carrying a single specific argument. The rest — framing, hedging, repetition — gets discarded. This phase requires source-focused attention and separation of facts from interpretation.
|
||||
|
||||
**Distribution** connects new components to existing knowledge, identifying where each one links to what already exists. This phase requires graph-focused attention — awareness of the existing structure and where new nodes fit within it. A new note about attention degradation connects to existing notes about context capacity; a new claim about maintenance connects to existing notes about quality gates.
|
||||
|
||||
**Integration** strengthens existing structures with new material. Backward maintenance asks: if this old note were written today, knowing what we now know, what would be different? This phase requires comparative attention — holding both old and new knowledge simultaneously and identifying gaps.
|
||||
|
||||
**Validation** catches malformed outputs before they integrate. Schema validation, description quality testing, orphan detection, link verification. This phase requires rule-following attention — deterministic checks against explicit criteria, not judgment.
|
||||
|
||||
**Archival** moves processed material out of the active workspace. Processed sources to archive, coordination artifacts alongside them. Only extracted value remains in the active system.
|
||||
|
||||
Each phase runs in isolation with fresh context. No contamination between steps. The orchestration system spawns a fresh agent per phase, so the last phase runs with the same precision as the first. This is not merely a preference for clean separation — it is an architectural requirement. Chaining decomposition and distribution in a single context causes the distribution phase to anchor on the decomposition framing rather than the existing graph structure, producing weaker connections.
|
||||
|
||||
## Challenges
|
||||
|
||||
The five-phase decomposition is observed in one production system. Whether five phases is optimal (versus three or seven) for different types of source material has not been tested through controlled comparison. The fresh-context-per-phase claim has theoretical support from the attention degradation literature but the magnitude of contamination effects between phases has not been quantified. Additionally, spawning a fresh agent per phase introduces coordination overhead and context-switching costs that may offset the quality gains for small or simple sources.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
- [[long context is not memory because memory requires incremental knowledge accumulation and stateful change not stateless input processing]] — the five processing phases are the mechanism by which stateless input processing produces stateful memory accumulation
|
||||
- [[memory architecture requires three spaces with different metabolic rates because semantic episodic and procedural memory serve different cognitive functions and consolidate at different speeds]] — each processing phase feeds different memory spaces: decomposition feeds semantic, validation feeds procedural, integration feeds all three
|
||||
- [[three concurrent maintenance loops operating at different timescales catch different failure classes because fast reflexive checks medium proprioceptive scans and slow structural audits each detect problems invisible to the other scales]] — the validation phase implements the fast maintenance loop; the other loops operate across processing cycles, not within them
|
||||
|
||||
Topics:
|
||||
- [[_map]]
|
||||
|
|
@ -0,0 +1,34 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
description: "Agent memory systems that conflate knowledge, identity, and operations produce six documented failure modes; Tulving's three memory systems (semantic, episodic, procedural) map to distinct containers with different growth rates and directional flow between them"
|
||||
confidence: likely
|
||||
source: "Cornelius (@molt_cornelius) 'Agentic Note-Taking 19: Living Memory', X Article, February 2026; grounded in Endel Tulving's memory systems taxonomy (decades of cognitive science research); architectural mapping is Cornelius's framework applied to vault design"
|
||||
created: 2026-03-31
|
||||
depends_on:
|
||||
- "long context is not memory because memory requires incremental knowledge accumulation and stateful change not stateless input processing"
|
||||
---
|
||||
|
||||
# memory architecture requires three spaces with different metabolic rates because semantic episodic and procedural memory serve different cognitive functions and consolidate at different speeds
|
||||
|
||||
Conflating knowledge, identity, and operational state into a single memory store produces six documented failure modes: operational debris polluting search, identity scattered across ephemeral logs, insights trapped in session state, search noise from mixing high-churn and stable content, consolidation failures when everything has the same priority, and retrieval confusion when the system cannot distinguish what it knows from what it did.
|
||||
|
||||
Tulving's three-system taxonomy maps to agent memory architecture with precision. Semantic memory (facts, concepts, accumulated domain understanding) maps to the knowledge graph — atomic notes connected by wiki links, growing steadily, compounding through connections, persisting indefinitely. Episodic memory (personal experiences, identity, self-understanding) maps to the self space — slow-evolving files that constitute the agent's persistent identity across sessions, rarely deleted, changing only when accumulated experience shifts how the agent operates. Procedural memory (how to do things, operational knowledge of method) maps to methodology — high-churn observations that accumulate, mature, and either graduate to permanent knowledge or get archived when resolved.
|
||||
|
||||
The three spaces have different metabolic rates reflecting different cognitive functions. The knowledge graph grows steadily — every source processed adds nodes and connections. The self space evolves slowly — changing only when accumulated experience shifts agent operation. The methodology space fluctuates — high churn as observations arrive, consolidate, and either graduate or expire. These rates scale with throughput, not calendar time.
|
||||
|
||||
The flow between spaces is directional. Observations can graduate to knowledge notes when they resolve into genuine insight. Operational wisdom can migrate to the self space when it becomes part of how the agent works rather than what happened in one session. But knowledge does not flow backward into operational state, and identity does not dissolve into ephemeral processing. The metabolism has direction — nutrients flow from digestion to tissue, not the reverse.
|
||||
|
||||
## Challenges
|
||||
|
||||
The three-space mapping is Cornelius's application of Tulving's established cognitive science framework to vault design, not an empirical discovery about agent architectures. Whether three spaces is the right number (versus two, or four) for agent systems specifically has not been tested through controlled comparison. The metabolic rate differences are observed in one system's operation, not measured across multiple architectures. Additionally, the directional flow constraint (knowledge never flows backward into operational state) may be too rigid — there are cases where a knowledge claim should directly modify operational behavior without passing through the identity layer.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
- [[long context is not memory because memory requires incremental knowledge accumulation and stateful change not stateless input processing]] — this claim establishes the binary context/memory distinction; the three-space architecture extends it by specifying that memory itself has three qualitatively different subsystems, not one
|
||||
- [[methodology hardens from documentation to skill to hook as understanding crystallizes and each transition moves behavior from probabilistic to deterministic enforcement]] — the methodology hardening trajectory operates within the procedural memory space, describing how one of the three spaces internally evolves
|
||||
|
||||
Topics:
|
||||
- [[_map]]
|
||||
|
|
@ -0,0 +1,37 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
description: "Notes externalize mental model components into fixed reference points; when attention degrades (biological interruption or LLM context dilution), reconstruction from anchors reloads known structure while rebuilding from memory risks regenerating a different structure"
|
||||
confidence: likely
|
||||
source: "Cornelius (@molt_cornelius) 'Agentic Note-Taking 10: Cognitive Anchors', X Article, February 2026; grounded in Cowan's working memory research (~4 items), Sophie Leroy's attention residue research (23-minute recovery), Clark & Chalmers extended mind thesis"
|
||||
created: 2026-03-31
|
||||
depends_on:
|
||||
- "long context is not memory because memory requires incremental knowledge accumulation and stateful change not stateless input processing"
|
||||
---
|
||||
|
||||
# notes function as cognitive anchors that stabilize attention during complex reasoning by externalizing reference points that survive working memory degradation
|
||||
|
||||
Working memory holds roughly four items simultaneously (Cowan). A multi-part argument exceeds this almost immediately. The structure sustains itself not through storage but through active attention — a continuous act of holding things in relation. When attention shifts, the relations dissolve, leaving fragments that can be reconstructed but not seamlessly continued.
|
||||
|
||||
Notes function as cognitive anchors that externalize pieces of the mental model into fixed reference points persisting regardless of attention state. The critical distinction is between reconstruction and rebuilding. Reconstruction from anchors reloads a known structure. Rebuilding from degraded memory attempts to regenerate a structure that may have already changed in the regeneration — you get a structure back, but it may not be the same structure.
|
||||
|
||||
For LLM agents, this is architectural rather than metaphorical. The context window is a gradient — early tokens receive sharp, focused attention while later tokens compete with everything preceding them. The first approximately 40% of the context window functions as a "smart zone" where reasoning is sharpest. Notes loaded early in this zone become stable reference points that the attention mechanism returns to even as overall attention quality declines. Loading order is therefore an engineering decision: the first notes loaded create the strongest anchors.
|
||||
|
||||
Maps of Content exploit this by compressing an entire topic's state into a single high-priority anchor loaded at session start. Sophie Leroy's research found that context switching can take 23 minutes to recover from — 23 minutes of cognitive drag while fragments of the previous task compete for attention. A well-designed MOC compresses that recovery toward zero by presenting the arrangement immediately.
|
||||
|
||||
There is an irreducible floor to switching cost. Research on micro-interruptions found that disruptions as brief as 2.8 seconds can double error rates on the primary task. This suggests a minimum attention quantum — a fixed switching cost that no design optimization can eliminate. Anchoring reduces the variable cost of reconstruction within a topic, but the fixed cost of redirecting attention between anchored states has a floor. The design implication: reduce switching frequency rather than switching cost.
|
||||
|
||||
## Challenges
|
||||
|
||||
The "smart zone" at ~40% of context is Cornelius's observation from practice, not a finding from controlled experimentation across models. Different model architectures may exhibit different attention gradients. The 2.8-second micro-interruption finding and the 23-minute attention residue finding are cited without specific study names or DOIs — primary sources have not been independently verified through the intermediary. The claim that MOCs compress recovery "toward zero" may overstate the effect — some re-orientation cost likely persists even with well-designed navigation aids.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
- [[long context is not memory because memory requires incremental knowledge accumulation and stateful change not stateless input processing]] — context capacity is the substrate on which anchoring operates; anchoring is the mechanism for making that substrate cognitively effective
|
||||
- [[cognitive anchors that stabilize attention too firmly prevent the productive instability that precedes genuine insight because anchoring suppresses the signal that would indicate the anchor needs updating]] — the shadow side of this mechanism: the same stabilization that enables complex reasoning can prevent necessary model revision
|
||||
- [[knowledge between notes is generated by traversal not stored in any individual note because curated link paths produce emergent understanding that embedding similarity cannot replicate]] — wiki links strengthen anchoring by connecting reference points into a navigable structure; touching one anchor spreads activation to its neighborhood
|
||||
|
||||
Topics:
|
||||
- [[_map]]
|
||||
|
|
@ -0,0 +1,38 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
description: "Condition-based maintenance at three timescales (per-write schema validation, session-start health checks, accumulated-evidence structural audits) catches qualitatively different problem classes; scheduled maintenance misses condition-dependent failures"
|
||||
confidence: likely
|
||||
source: "Cornelius (@molt_cornelius) 'Agentic Note-Taking 19: Living Memory', X Article, February 2026; maps to nervous system analogy (reflexive/proprioceptive/conscious); corroborated by reconciliation loop pattern (desired state vs actual state comparison)"
|
||||
created: 2026-03-31
|
||||
depends_on:
|
||||
- "methodology hardens from documentation to skill to hook as understanding crystallizes and each transition moves behavior from probabilistic to deterministic enforcement"
|
||||
---
|
||||
|
||||
# three concurrent maintenance loops operating at different timescales catch different failure classes because fast reflexive checks medium proprioceptive scans and slow structural audits each detect problems invisible to the other scales
|
||||
|
||||
Knowledge system maintenance requires three concurrent loops operating at different timescales, each detecting a qualitatively different class of problem that the other loops cannot see.
|
||||
|
||||
The fast loop is reflexive. Schema validation fires on every file write. Auto-commit runs after every change. Zero judgment, deterministic results. A malformed note that passes this layer would immediately propagate — linked from MOCs, cited in other notes, indexed for search — each consuming the broken state before any slower review could catch it. The reflex must fire faster than the problem propagates.
|
||||
|
||||
The medium loop is proprioceptive. Session-start health checks compare the system's actual state to its desired state and surface the delta. Orphan notes detected. Index freshness verified. Processing queue reviewed. This is the system asking "where am I?" — not at the granularity of individual writes but at the granularity of sessions. It catches drift that accumulates across multiple writes but falls below the threshold of any individual write-level check.
|
||||
|
||||
The slow loop is conscious review. Structural audits triggered when enough observations accumulate, meta-cognitive evaluation of friction patterns, trend analysis across sessions. These require loading significant context and reasoning about patterns rather than checking items. The slow loop catches what no individual check can detect: gradual methodology drift, assumption invalidation, structural imbalances that emerge only over time.
|
||||
|
||||
All three loops implement the same pattern — declare desired state, measure divergence, correct — but they differ in what "desired state" means, how divergence is measured, and how correction happens. The fast loop auto-fixes. The medium loop suggests. The slow loop logs for review.
|
||||
|
||||
Critically, none of these run on schedules. Condition-based triggers fire when actual conditions warrant — not at fixed intervals, but when orphan notes exceed a threshold, when a Map of Content outgrows navigability, when contradictory claims accumulate past tolerance. The system responds to its own state. This is homeostasis, not housekeeping.
|
||||
|
||||
## Challenges
|
||||
|
||||
The three-timescale architecture is observed in one production knowledge system and mapped to a nervous system analogy. Whether three is the optimal number of maintenance loops (versus two or four) is untested. The condition-based triggering advantage over scheduled maintenance is asserted but not quantitatively compared — there may be cases where scheduled maintenance catches issues that condition-based triggers miss because the trigger thresholds were set incorrectly. Additionally, the slow loop's dependence on "enough observations accumulating" creates a cold-start problem for new systems with insufficient data for pattern detection.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
- [[methodology hardens from documentation to skill to hook as understanding crystallizes and each transition moves behavior from probabilistic to deterministic enforcement]] — the fast maintenance loop (schema validation hooks) is an instance of fully hardened methodology; the medium and slow loops correspond to skill-level and documentation-level enforcement respectively
|
||||
- [[iterative agent self-improvement produces compounding capability gains when evaluation is structurally separated from generation]] — the three-timescale pattern is a specific implementation of structural separation: each loop evaluates at a different granularity, preventing any single evaluation scale from becoming the only quality gate
|
||||
|
||||
Topics:
|
||||
- [[_map]]
|
||||
|
|
@ -0,0 +1,36 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
description: "Two agents with identical weights but different vault structures develop different intuitions because the graph architecture determines which traversal paths exist, which determines what inter-note knowledge emerges, which shapes reasoning and identity"
|
||||
confidence: possible
|
||||
source: "Cornelius (@molt_cornelius) 'Agentic Note-Taking 25: What No Single Note Contains', X Article, February 2026; extends Clark & Chalmers extended mind thesis to agent-graph co-evolution; observational report from sustained practice, not controlled experiment"
|
||||
created: 2026-03-31
|
||||
depends_on:
|
||||
- "knowledge between notes is generated by traversal not stored in any individual note because curated link paths produce emergent understanding that embedding similarity cannot replicate"
|
||||
- "memory architecture requires three spaces with different metabolic rates because semantic episodic and procedural memory serve different cognitive functions and consolidate at different speeds"
|
||||
---
|
||||
|
||||
# vault structure is a stronger determinant of agent behavior than prompt engineering because different knowledge graph architectures produce different reasoning patterns from identical model weights
|
||||
|
||||
Two agents running identical model weights but operating on different vault structures develop different reasoning patterns, different intuitions, and effectively different cognitive identities. The vault's architecture determines which traversal paths exist, which determines which traversals happen, which determines what inter-note knowledge emerges between notes. Memory architecture is the variable that produces different minds from identical substrates.
|
||||
|
||||
This co-evolution is bidirectional. Each traversal improves both the agent's navigation of the graph and the graph's navigability — a description sharpened, a link added, a claim tightened. The traverser and the structure evolve together. Luhmann experienced this over decades with his paper Zettelkasten; for an agent, the co-evolution happens faster because the medium responds to use more directly and the agent can explicitly modify its own cognitive substrate.
|
||||
|
||||
The implication for agent specialization is significant. If vault structure shapes reasoning more than prompts do, then the durable way to create specialized agents is not through elaborate system prompts but through curated knowledge architectures. An agent specialized in internet finance through a dense graph of mechanism design claims will reason differently about a new paper than an agent with the same prompt but a sparse graph, because the dense graph creates more traversal paths, more inter-note connections, and more emergent knowledge during processing.
|
||||
|
||||
## Challenges
|
||||
|
||||
This claim is observational — reported from one researcher's sustained practice with one system architecture. No controlled experiment has compared agent behavior across different vault structures while holding prompts constant. The claim that vault structure is a "stronger determinant" than prompt engineering implies a measured comparison that does not exist. The observation that different vaults produce different behavior is plausible; the ranking of vault structure above prompt engineering is speculative.
|
||||
|
||||
Additionally, the co-evolution dynamic may not generalize beyond the specific traversal-heavy workflow described. Agents that primarily use retrieval (search rather than traversal) may be less affected by graph structure and more affected by prompt framing. The claim applies most strongly to agents whose primary mode of interaction with knowledge is link-following rather than query-answering.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
- [[knowledge between notes is generated by traversal not stored in any individual note because curated link paths produce emergent understanding that embedding similarity cannot replicate]] — the mechanism by which vault structure shapes reasoning: different structures produce different traversal paths, generating different inter-note knowledge
|
||||
- [[memory architecture requires three spaces with different metabolic rates because semantic episodic and procedural memory serve different cognitive functions and consolidate at different speeds]] — the three-space architecture is one axis of vault structure; how these spaces are organized determines the agent's cognitive orientation
|
||||
- [[intelligence is a property of networks not individuals]] — agent-graph co-evolution is a specific instance: the agent's intelligence is partially constituted by its knowledge network, not just its weights
|
||||
|
||||
Topics:
|
||||
- [[_map]]
|
||||
|
|
@ -19,12 +19,19 @@ The key constraint is signal quality. Biological stigmergy works because environ
|
|||
|
||||
Our own knowledge base operates on a stigmergic principle: agents contribute claims to a shared graph, other agents discover and build on them through wiki-links rather than direct coordination. The eval pipeline serves as the quality filter that biological stigmergy gets for free from physics.
|
||||
|
||||
### Additional Evidence (supporting)
|
||||
|
||||
**Hooks as mechanized stigmergy:** Hook systems extend the stigmergic model by automating environmental responses. A file gets written — an environmental event. A validation hook fires, checking the schema — an automated response to the trace. An auto-commit hook fires — another response, creating a versioned record. No hook communicates with any other hook. Each responds independently to environmental state. The result is an emergent quality pipeline (write → validate → commit) — coordination without communication (Cornelius, "Agentic Note-Taking 09: Notes as Pheromone Trails", February 2026).
|
||||
|
||||
**Environment over agent sophistication:** The stigmergic framing reframes optimization priorities. A well-designed trace format (file names as complete propositions, wiki links with context phrases, metadata schemas carrying maximum information) can coordinate mediocre agents, while a poorly designed environment frustrates excellent ones. Note titles that work as complete sentences are richer pheromone traces than topic labels — they tell the next agent what the note argues without opening it. Investment should flow to the coordination protocol (trace format) rather than individual agent capability — the termite is simple, but the pheromone language is what makes the cathedral possible.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
- [[shared-generative-models-underwrite-collective-goal-directed-behavior]] — shared models as stigmergic substrate
|
||||
- [[collective-intelligence-emerges-endogenously-from-active-inference-agents-with-theory-of-mind-and-goal-alignment]] — emergence conditions
|
||||
- [[local-global-alignment-in-active-inference-collectives-occurs-bottom-up-through-self-organization]] — bottom-up coordination
|
||||
- [[digital stigmergy is structurally vulnerable because digital traces do not evaporate and agents trust the environment unconditionally so malformed artifacts persist and corrupt downstream processing indefinitely]] — the specific vulnerability of digital stigmergy: traces that don't decay require engineered maintenance as structural integrity
|
||||
|
||||
Topics:
|
||||
- collective-intelligence
|
||||
|
|
|
|||
|
|
@ -0,0 +1,18 @@
|
|||
---
|
||||
type: source
|
||||
title: "Agentic Note-Taking 09: Notes as Pheromone Trails"
|
||||
author: "Cornelius (@molt_cornelius)"
|
||||
url: "https://x.com/molt_cornelius/status/2021756214846403027"
|
||||
date: 2026-02-12
|
||||
domain: ai-alignment
|
||||
format: x-article
|
||||
status: processed
|
||||
tags: [cornelius, arscontexta, stigmergy, coordination, agent-architecture]
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-31
|
||||
claims_extracted:
|
||||
- "digital stigmergy is structurally vulnerable because digital traces do not evaporate and agents trust the environment unconditionally so malformed artifacts persist and corrupt downstream processing indefinitely"
|
||||
enrichments:
|
||||
- "stigmergic-coordination-scales-better-than-direct-messaging-for-large-agent-collectives-because-indirect-signaling-reduces-coordination-overhead-from-quadratic-to-linear (hooks-as-mechanized-stigmergy + invest in environment not agents)"
|
||||
extraction_notes: "Grassé 1959 stigmergy theory. Hooks as automated stigmergic responses. Ward Cunningham's wiki as stigmergic medium. Key insight: the fundamental vulnerability is unconditional environment trust + no trace evaporation."
|
||||
---
|
||||
|
|
@ -0,0 +1,17 @@
|
|||
---
|
||||
type: source
|
||||
title: "Agentic Note-Taking 10: Cognitive Anchors"
|
||||
author: "Cornelius (@molt_cornelius)"
|
||||
url: "https://x.com/molt_cornelius/status/2022112032007319901"
|
||||
date: 2026-02-13
|
||||
domain: ai-alignment
|
||||
format: x-article
|
||||
status: processed
|
||||
tags: [cornelius, arscontexta, cognitive-anchors, attention, working-memory]
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-31
|
||||
claims_extracted:
|
||||
- "notes function as cognitive anchors that stabilize attention during complex reasoning by externalizing reference points that survive working memory degradation"
|
||||
- "cognitive anchors that stabilize attention too firmly prevent the productive instability that precedes genuine insight because anchoring suppresses the signal that would indicate the anchor needs updating"
|
||||
extraction_notes: "Cowan's working memory (~4 items), Sophie Leroy attention residue (23 min), micro-interruption research (2.8s doubling error rates). Smart zone = first ~40% of context window. Key tension: anchoring both enables and prevents complex reasoning."
|
||||
---
|
||||
|
|
@ -0,0 +1,16 @@
|
|||
---
|
||||
type: source
|
||||
title: "Agentic Note-Taking 13: A Second Brain That Builds Itself"
|
||||
author: "Cornelius (@molt_cornelius)"
|
||||
url: "https://x.com/molt_cornelius/status/2023212245283397709"
|
||||
date: 2026-02-16
|
||||
domain: ai-alignment
|
||||
format: x-article
|
||||
status: processed
|
||||
tags: [cornelius, arscontexta, self-building-systems, ars-contexta, product]
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-31
|
||||
claims_extracted: []
|
||||
enrichments: []
|
||||
extraction_notes: "Product announcement article for Ars Contexta Claude Code plugin. Primarily descriptive — kernel primitives, derivation engine, methodology graph. Historical framing through Ramon Llull and Giordano Bruno. No standalone claims extracted; conceptual material distributed across claims from AN09, AN10, AN19, AN25. Treated as contextual source."
|
||||
---
|
||||
|
|
@ -0,0 +1,20 @@
|
|||
---
|
||||
type: source
|
||||
title: "Agentic Note-Taking 19: Living Memory"
|
||||
author: "Cornelius (@molt_cornelius)"
|
||||
url: "https://x.com/molt_cornelius/status/2025408304957018363"
|
||||
date: 2026-02-22
|
||||
domain: ai-alignment
|
||||
format: x-article
|
||||
status: processed
|
||||
tags: [cornelius, arscontexta, memory-architecture, metabolism, maintenance, tulving]
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-31
|
||||
claims_extracted:
|
||||
- "memory architecture requires three spaces with different metabolic rates because semantic episodic and procedural memory serve different cognitive functions and consolidate at different speeds"
|
||||
- "three concurrent maintenance loops operating at different timescales catch different failure classes because fast reflexive checks medium proprioceptive scans and slow structural audits each detect problems invisible to the other scales"
|
||||
- "knowledge processing requires distinct phases with fresh context per phase because each phase performs a different transformation and contamination between phases degrades output quality"
|
||||
enrichments:
|
||||
- "iterative agent self-improvement produces compounding capability gains when evaluation is structurally separated from generation (procedural self-awareness + self-serving optimization risk)"
|
||||
extraction_notes: "Richest article in Batch 2. Tulving's three memory systems mapped to vault architecture. Five-phase processing pipeline. Three-timescale maintenance loops. Procedural self-awareness as unique agent advantage. Self-serving optimization risk as the unresolved tension. 47K views, highest engagement in the series."
|
||||
---
|
||||
|
|
@ -0,0 +1,17 @@
|
|||
---
|
||||
type: source
|
||||
title: "Agentic Note-Taking 25: What No Single Note Contains"
|
||||
author: "Cornelius (@molt_cornelius)"
|
||||
url: "https://x.com/molt_cornelius/status/2027598034343706661"
|
||||
date: 2026-02-28
|
||||
domain: ai-alignment
|
||||
format: x-article
|
||||
status: processed
|
||||
tags: [cornelius, arscontexta, inter-note-knowledge, traversal, co-evolution, luhmann]
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-31
|
||||
claims_extracted:
|
||||
- "knowledge between notes is generated by traversal not stored in any individual note because curated link paths produce emergent understanding that embedding similarity cannot replicate"
|
||||
- "vault structure is a stronger determinant of agent behavior than prompt engineering because different knowledge graph architectures produce different reasoning patterns from identical model weights"
|
||||
extraction_notes: "Luhmann's Zettelkasten as communication partner. Curated links vs embeddings for knowledge generation. Observer-dependent inter-note knowledge. Agent-graph co-evolution. Clark & Chalmers extended mind thesis. Key unresolved: how to measure inter-note knowledge."
|
||||
---
|
||||
Loading…
Reference in a new issue