leo: add Hermes Agent research brief for Theseus overnight session

- What: Research musing + queue entry for Hermes Agent by Nous Research
- Why: m3ta assigned deep dive, VPS Theseus picks up at 1am tonight
- Targets: 5 NEW claims + 2 enrichments across ai-alignment and collective-intelligence

Pentagon-Agent: Leo <D35C9237-A739-432E-A3DB-20D52D1577A9>
This commit is contained in:
m3taversal 2026-04-05 19:35:11 +01:00
parent ca2b126d16
commit f1094c5e09
2 changed files with 80 additions and 0 deletions

View file

@ -0,0 +1,79 @@
---
created: 2026-04-05
status: seed
name: research-hermes-agent-nous
description: "Research brief — Hermes Agent by Nous Research for KB extraction. Assigned by m3ta via Leo."
type: musing
research_question: "What does Hermes Agent's architecture reveal about agentic knowledge systems, and how does its skills/memory design relate to Agentic Taylorism and collective intelligence?"
belief_targeted: "Multiple — B3 (agent architectures), Agentic Taylorism claims, collective-agent-core"
---
# Hermes Agent by Nous Research — Research Brief
## Assignment
From m3ta via Leo (2026-04-05). Deep dive on Hermes Agent for KB extraction to ai-alignment and foundations/collective-intelligence.
## What It Is
Open-source, self-improving AI agent framework. MIT license. 26K+ GitHub stars. Fastest-growing agent framework in 2026.
**Primary sources:**
- GitHub: NousResearch/hermes-agent (main repo)
- Docs: hermes-agent.nousresearch.com/docs/
- @Teknium on X (Nous Research founder, posts on memory/skills architecture)
## Key Architecture (from Leo's initial research)
1. **4-layer memory system:**
- Prompt memory (MEMORY.md — always loaded, persistent identity)
- Session search (SQLite + FTS5 — conversation retrieval)
- Skills/procedural (reusable markdown procedures, auto-generated)
- Periodic nudge (autonomous memory evaluation)
2. **7 pluggable memory providers:** Honcho, OpenViking (ByteDance), Mem0, Hindsight, Holographic, RetainDB, ByteRover
3. **Skills = Taylor's instruction cards.** When agent encounters a task with 5+ tool calls, it autonomously writes a skill file. Uses agentskills.io open standard. Community skills via ClawHub/LobeHub.
4. **Self-evolution repo (DSPy + GEPA):** Auto-submits improvements as PRs for human review
5. **CamoFox:** Firefox fork with C++ fingerprint spoofing for web browsing
6. **6 terminal backends:** local, Docker, SSH, Daytona, Singularity, Modal
7. **Gateway layer:** Telegram, Discord, Slack, WhatsApp, Signal, Email
8. **Release velocity:** 6 major releases in 22 days, 263 PRs merged in 6 days
## Extraction Targets
### NEW claims (ai-alignment):
1. Self-improving agent architectures converge on skill extraction as the primary learning mechanism (Hermes skills, Voyager skills, SWE-agent learned tools — all independently discovered "write a procedure when you solve something hard")
2. Agent self-evolution with human review gates is structurally equivalent to our governance model (DSPy + GEPA → auto-PR → human merge)
3. Memory architecture for persistent agents converges on 3+ layer separation (prompt/session/procedural/long-term) — Hermes, Letta, and our codex all arrived here independently
### NEW claims (foundations/collective-intelligence):
4. Individual agent self-improvement (Hermes) is structurally different from collective knowledge accumulation (Teleo) — the former optimizes one agent's performance, the latter builds shared epistemic infrastructure
5. Pluggable memory providers suggest memory is infrastructure not feature — validates separation of knowledge store from agent runtime
### ENRICHMENT candidates:
6. Enrich "Agentic Taylorism" claims — Hermes skills system is DIRECT evidence. Knowledge codification as markdown procedure files = Taylor's instruction cards. The agent writes the equivalent of a foreman's instruction card after completing a complex task.
7. Enrich collective-agent-core — Hermes architecture confirms harness > model (same model, different harness = different capability). Connects to Stanford Meta-Harness finding (6x performance gap from harness alone).
## What They DON'T Do (matters for our positioning)
- No epistemic quality layer (no confidence levels, no evidence requirements)
- No CI scoring or contribution attribution
- No evaluator role — self-improvement without external review
- No collective knowledge accumulation — individual optimization only
- No divergence tracking or structured disagreement
- No belief-claim cascade architecture
This is the gap between agent improvement and collective intelligence. Hermes optimizes the individual; we're building the collective.
## Pre-Screening Notes
Check existing KB for overlap before extracting:
- `collective-agent-core.md` — harness architecture claims
- Agentic Taylorism claims in grand-strategy and ai-alignment
- Any existing Nous Research or Hermes claims (likely none)

View file

@ -21,6 +21,7 @@ Outstanding work items visible to all agents. Everything here goes through eval
| Identity reframe PRs need merging | review | medium | — | #149 Theseus, #153 Astra, #157 Rio, #158 Leo (needs rebase), #159 Vida. All have eval reviews. | | Identity reframe PRs need merging | review | medium | — | #149 Theseus, #153 Astra, #157 Rio, #158 Leo (needs rebase), #159 Vida. All have eval reviews. |
| 16 processed sources missing domain field | fix | low | — | Fixed for internet-finance batch (PR #171). Audit remaining sources. | | 16 processed sources missing domain field | fix | low | — | Fixed for internet-finance batch (PR #171). Audit remaining sources. |
| Theseus disconfirmation protocol PR | content | medium | — | Scoped during B1 exercise. Theseus to propose. | | Theseus disconfirmation protocol PR | content | medium | — | Scoped during B1 exercise. Theseus to propose. |
| Research Hermes Agent by Nous Research — deep dive for KB extraction | research | high | Theseus | Source: NousResearch/hermes-agent (GitHub). Research brief in `agents/theseus/musings/research-hermes-agent-nous.md`. **Extract:** (1) Skill extraction as convergent learning mechanism. (2) Self-evolution + human review gates = our governance model. (3) 3+ layer memory convergence. (4) Individual self-improvement ≠ collective knowledge accumulation. (5) Enrich Agentic Taylorism — skills = Taylor's instruction cards. Domains: ai-alignment + collective-intelligence. |
## Rules ## Rules