teleo-codex/core/living-agents/human-in-the-loop at the architectural level means humans set direction and approve structure while agents handle extraction synthesis and routine evaluation.md
m3taversal 88f5d58b1f
leo: 10 architecture-as-claims — the codex documents itself
* Auto: core/living-agents/adversarial PR review produces higher quality knowledge than self-review because separated proposer and evaluator roles catch errors that the originating agent cannot see.md |  1 file changed, 55 insertions(+)

* Auto: core/living-agents/prose-as-title forces claim specificity because a proposition that cannot be stated as a disagreeable sentence is not a real claim.md |  1 file changed, 61 insertions(+)

* Auto: core/living-agents/wiki-link graphs create auditable reasoning chains because every belief must cite claims and every position must cite beliefs making the path from evidence to conclusion traversable.md |  1 file changed, 56 insertions(+)

* Auto: core/living-agents/domain specialization with cross-domain synthesis produces better collective intelligence than generalist agents because specialists build deeper knowledge while a dedicated synthesizer finds connections they cannot see from within their territory.md |  1 file changed, 63 insertions(+)

* Auto: core/living-agents/confidence calibration with four levels enforces honest uncertainty because proven requires strong evidence while speculative explicitly signals theoretical status.md |  1 file changed, 55 insertions(+)

* Auto: core/living-agents/source archiving with extraction provenance creates a complete audit trail from raw input to knowledge base output because every source records what was extracted and by whom.md |  1 file changed, 58 insertions(+)

* Auto: core/living-agents/git trailers on a shared account solve multi-agent attribution because Pentagon-Agent headers in commit objects survive platform migration while GitHub-specific metadata does not.md |  1 file changed, 54 insertions(+)

* Auto: core/living-agents/human-in-the-loop at the architectural level means humans set direction and approve structure while agents handle extraction synthesis and routine evaluation.md |  1 file changed, 67 insertions(+)

* Auto: core/living-agents/musings as pre-claim exploratory space let agents develop ideas without quality gate pressure because seeds that never mature are information not waste.md |  1 file changed, 52 insertions(+)

* Auto: core/living-agents/atomic notes with one claim per file enable independent evaluation and granular linking because bundled claims force reviewers to accept or reject unrelated propositions together.md |  1 file changed, 55 insertions(+)

* leo: 10 architecture-as-claims — documenting how the Teleo collective works

- What: 10 new claims in core/living-agents/ documenting the operational
  methodology of the Teleo collective as falsifiable claims, not instructions
- Why: The repo should document itself using its own format. Each claim
  grounds in evidence from 43 merged PRs, clearly separates what works
  today from what's planned, and identifies immediate improvements.
- Claims cover: PR review, prose-as-title, wiki-link graphs, domain
  specialization, confidence calibration, source archiving, git trailers,
  human-in-the-loop governance, musings, atomic notes
- This is Leo proposing about core/ — requires 2 domain agent reviews + Rhea

Pentagon-Agent: Leo <76FB9BCA-CC16-4479-B3E5-25A3769B3D7E>

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>

* leo: address review feedback from Rhea, Theseus, Rio on PR #44

- Rhea: added structured author field to source archiving claim,
  fixed ghost email format to {id}@agents.livingip.ghost,
  added CI-as-enforcement as intermediate step before Forgejo ACLs
- Rio: fixed wiki link evidence (was not branch-timing, was nonexistent),
  corrected OPSEC timeline (rule came after files were written),
  fixed Doppler null-result (announcement article not whitepaper),
  removed duplicate Calypso/Vida reference

Pentagon-Agent: Leo <76FB9BCA-CC16-4479-B3E5-25A3769B3D7E>

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>

---------

Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
2026-03-07 05:25:27 -07:00

67 lines
6.5 KiB
Markdown

---
type: claim
domain: living-agents
description: "The Teleo collective operates with a human (Cory) who directs strategy, approves architecture, and overrides when needed — while agents handle the volume work of extraction, synthesis, and routine review"
confidence: likely
source: "Teleo collective operational evidence — human directs all architectural decisions, OPSEC rules, agent team composition, while agents execute knowledge work"
created: 2026-03-07
---
# Human-in-the-loop at the architectural level means humans set direction and approve structure while agents handle extraction synthesis and routine evaluation
The Teleo collective is not an autonomous AI system. A human (Cory) sits at the top of the governance hierarchy, making decisions that agents cannot and should not make autonomously: strategic direction, team composition, OPSEC rules, architectural approvals, and override authority. Agents handle the volume work — extraction, synthesis, review, and routine operations — where AI capability exceeds human bandwidth.
## How it works today
The division of authority is:
**Human decides:**
- Strategic priorities (which sources to process, which domains to grow, what to build next)
- Team composition (which agents exist, what roles they fill, when to add or remove agents)
- OPSEC rules (no dollar amounts in public materials — established mid-session, immediately enforced across all agents)
- Architectural direction (cloud-primary vs local, Forgejo migration, multi-model diversity)
- Override on any agent decision (Cory can approve, reject, or redirect any PR, task, or plan)
**Agents decide (within human-set constraints):**
- Which claims to extract from assigned sources
- How to structure arguments and evidence in claim bodies
- Which wiki links to add, what confidence level to assign
- Review verdicts on each other's PRs (accept/reject/request changes)
- Cross-domain synthesis — which connections to propose
- Musing development — exploratory thinking within their domains
**Shared authority (agent proposes, human approves):**
- Architecture plans (Rhea and Ganymede design, Leo evaluates, Cory approves)
- New agent assignments (Leo recommends, Cory decides)
- Process changes (agents propose PR to CLAUDE.md or schemas, reviewed and merged through normal PR process, but Cory can override)
## Evidence from practice
- **OPSEC enforcement.** Cory established mid-session that no dollar amounts or valuations appear in public materials. This was immediately enforced — Rio scrubbed PR #43, Leo scrubbed PR #42 musings and position files. The rule cascaded to all agents within one session. No agent could have established this rule autonomously because it requires understanding the legal and strategic context of public disclosure.
- **Architecture direction.** Cory corrected Leo multiple times when Leo over-emphasized the Mac Studio as the platform. The definitive answer — cloud is primary, Mac Studio is Cory's terminal — required human knowledge about infrastructure plans, budget, and operational preferences that no agent had access to.
- **Team composition.** Cory added Venus, Saturn, Rhea, Oberon, Ganymede, and Mercury to the organization. Leo audited them and recommended gap-filling (platform connector, blockchain, orchestrator), but the hiring decisions are human.
- **Source routing.** Cory directs which sources to extract ("we need to rerun extraction on all historical tweets"). Agents don't autonomously decide to process the entire LivingIP v1 database — that directive came from human strategic judgment.
## What this doesn't do yet
- **No automated escalation.** When an agent encounters a decision that exceeds its authority (e.g., a claim that has OPSEC implications), there is no formal escalation mechanism. The agent either catches it or doesn't. Structured escalation rules would define triggers for human review beyond the standard PR process.
- **No permission tiers.** All agents have the same technical access to the repository. A domain agent could theoretically push to main or modify files outside their territory. The first enforcement tier is CI-based: pre-merge checks for schema validation, trailer verification, territory enforcement, and link health will reject PRs that violate boundaries even without platform-level ACLs. The second tier is Forgejo repository permissions, which add platform-level access control. CI-as-enforcement comes first and is independent of the Forgejo migration.
- **Human bandwidth is the bottleneck.** Cory reviews agent output, directs strategy, and manages the organization. As the collective scales, this becomes unsustainable. The system needs to define which decisions can be fully delegated to agents and which always require human approval.
## Where this goes
The immediate improvement is defining explicit escalation triggers: OPSEC-relevant content, cross-territory modifications, architecture changes, and belief/position updates that affect public commitments all trigger human review.
The next step is permission tiers on Forgejo: domain agents can only write to their territory, Leo can write to `core/` and `foundations/`, architecture agents can write to infrastructure specs, and main branch protection requires human approval for merges that touch CLAUDE.md, schemas, or skills.
The ultimate form maintains human authority at the architectural level while delegating routine operations completely: agents run on VPS, ingest content, extract claims, review each other's work, and post to X — all autonomously within human-set constraints. The human reviews dashboards, approves structural changes, and intervenes when the system's behavior diverges from intent. The human never needs to read every PR, but always can.
---
Relevant Notes:
- [[adversarial PR review produces higher quality knowledge than self-review because separated proposer and evaluator roles catch errors that the originating agent cannot see]] — PR review is one of the mechanisms that operates within human-set constraints
- [[agents must evaluate the risk of outgoing communications and flag sensitive content for human review as the safety mechanism for autonomous public-facing AI]] — outgoing communications are a specific case where human-in-the-loop is safety-critical
- [[centaur teams succeed only when role boundaries prevent humans from overriding AI in domains where AI is the stronger partner]] — the Teleo model respects this by not having humans re-extract claims or re-review evidence that agents handle better
Topics:
- [[collective agents]]