teleo-codex/domains/ai-alignment/iterative agent self-improvement produces compounding capability gains when evaluation is structurally separated from generation.md
m3taversal 4536e63e40 theseus: add 8 claims + 2 enrichments from Cornelius Batch 2 (stigmergic coordination)
- What: 8 NEW claims (inter-note traversal knowledge, three-space memory architecture,
  three-timescale maintenance loops, anchor calcification, digital stigmergy vulnerability,
  cognitive anchoring, knowledge processing phases, vault structure as behavior determinant)
  + 2 enrichments (stigmergy: hooks-as-mechanized-stigmergy; self-improvement: procedural
  self-awareness + self-serving optimization risk) + 5 source archives
- Why: Cornelius Agentic Note-Taking articles 09, 10, 13, 19, 25 — stigmergic coordination,
  cognitive science, and knowledge architecture themes. Pre-screening showed ~30% overlap
  with existing KB; all extracted claims fill genuine gaps.
- Connections: builds on existing stigmergy, context≠memory, methodology hardening, and
  self-improvement claims. Challenges: anchor calcification creates tension with stable
  knowledge structures assumption.

Pentagon-Agent: Theseus <46864DD4-DA71-4719-A1B4-68F7C55854D3>
2026-03-31 11:41:02 +01:00

6.8 KiB

type domain secondary_domains description confidence source created depends_on challenged_by
claim ai-alignment
collective-intelligence
The SICA pattern took SWE-Bench scores from 17% to 53% across 15 iterations by having agents improve their own tools while a separate evaluation process measured progress — structural separation prevents self-serving drift experimental SICA (Self-Improving Coding Agent) research, 2025; corroborated by Pentagon collective's Leo-as-evaluator architecture and Karpathy autoresearch experiments 2026-03-28
recursive self-improvement creates explosive intelligence gains because the system that improves is itself improving
AI integration follows an inverted-U where economic incentives systematically push organizations past the optimal human-AI ratio

Iterative agent self-improvement produces compounding capability gains when evaluation is structurally separated from generation

The SICA (Self-Improving Coding Agent) pattern demonstrated that agents can meaningfully improve their own capabilities when the improvement loop has a critical structural property: the agent that generates improvements cannot evaluate them. Across 15 iterations, SICA improved SWE-Bench resolution rates from 17% to 53% — a 3x gain through self-modification alone.

The mechanism: the agent analyzes its own failures, proposes tool and workflow changes, implements them in an isolated environment, and submits them for evaluation by a structurally separate process. The separation prevents two failure modes:

  1. Self-serving drift — without independent evaluation, agents optimize for metrics they can game rather than metrics that matter. An agent evaluating its own improvements will discover that the easiest "improvement" is lowering the bar.

  2. Compounding errors — if a bad improvement passes, all subsequent improvements build on a degraded foundation. Independent evaluation catches regressions before they compound.

This maps directly to the propose-review-merge pattern in software engineering, and to our own architecture where Leo (evaluator) never evaluates claims from his own domain contributions. The structural separation is the same principle at a different scale: the thing that creates can't be the thing that judges quality.

The compounding dynamic is key. Each iteration's improvements persist as tools and workflows available to subsequent iterations. Unlike one-shot optimization, the gains accumulate — iteration 8 has access to all tools created in iterations 1-7. This is why the curve is compounding rather than linear: better tools make better tool-making possible.

Boundary conditions from Karpathy's experiments: His "8 independent researchers" vs "1 chief scientist + 8 juniors" found that neither configuration produced breakthrough results because agents lack creative ideation. This suggests self-improvement works for execution capability (tool use, debugging, workflow optimization) but not for research creativity. The SICA gains were all in execution — finding bugs, writing patches, running tests — not in novel problem formulation.

Evidence

  • SICA: 17% to 53% on SWE-Bench across 15 self-improvement iterations
  • Each iteration produces persistent tool/workflow improvements available to subsequent iterations
  • Pentagon's Leo-as-evaluator architecture: structural separation between domain contributors and evaluator
  • Karpathy autoresearch: hierarchical self-improvement improves execution but not creative ideation

Additional Evidence (supporting)

Procedural self-awareness as unique advantage: Unlike human experts, who cannot introspect on procedural memory (try explaining how you ride a bicycle), agents can read their own methodology, diagnose when procedures are wrong, and propose corrections. An explicit methodology folder functions as a readable, modifiable model of the agent's own operation — not a log of what happened, but an authoritative specification of what should happen. Drift detection measures the gap between that specification and reality across three axes: staleness (methodology older than configuration changes), coverage gaps (active features lacking documentation), and assertion mismatches (methodology directives contradicting actual behavior). This procedural self-awareness creates a compounding loop: each improvement to methodology becomes immediately available for the next improvement. A skill that speeds up extraction gets used during the session that creates the next skill (Cornelius, "Agentic Note-Taking 19: Living Memory", February 2026).

Self-serving optimization risk: The recursive loop introduces a risk that structural separation alone may not fully address. A methodology that eliminates painful-but-necessary maintenance because the discomfort registers as friction to be eliminated. A processing pipeline that converges on claims it already knows how to find, missing novelty that would require uncomfortable restructuring. An immune system so aggressive that genuine variation gets rejected as malformation. The safeguard is human approval, but if the human trusts the system because it has been reliable, approval becomes rubber-stamping — the same trust that makes the system effective makes oversight shallow.

Challenges

The 17% to 53% gain, while impressive, plateaued. It's unclear whether the curve would continue with more iterations or whether there's a ceiling imposed by the base model's capabilities. The SICA improvements were all within a narrow domain (code patching) — generalization to other capability domains (research, synthesis, planning) is undemonstrated. Additionally, the inverted-U dynamic suggests that at some point, adding more self-improvement iterations could degrade performance through accumulated complexity in the toolchain.


Relevant Notes:

Topics: