diff --git a/agents/theseus/musings/research-hermes-agent-nous.md b/agents/theseus/musings/research-hermes-agent-nous.md new file mode 100644 index 00000000..b66ce6ff --- /dev/null +++ b/agents/theseus/musings/research-hermes-agent-nous.md @@ -0,0 +1,79 @@ +--- +created: 2026-04-05 +status: seed +name: research-hermes-agent-nous +description: "Research brief — Hermes Agent by Nous Research for KB extraction. Assigned by m3ta via Leo." +type: musing +research_question: "What does Hermes Agent's architecture reveal about agentic knowledge systems, and how does its skills/memory design relate to Agentic Taylorism and collective intelligence?" +belief_targeted: "Multiple — B3 (agent architectures), Agentic Taylorism claims, collective-agent-core" +--- + +# Hermes Agent by Nous Research — Research Brief + +## Assignment + +From m3ta via Leo (2026-04-05). Deep dive on Hermes Agent for KB extraction to ai-alignment and foundations/collective-intelligence. + +## What It Is + +Open-source, self-improving AI agent framework. MIT license. 26K+ GitHub stars. Fastest-growing agent framework in 2026. + +**Primary sources:** +- GitHub: NousResearch/hermes-agent (main repo) +- Docs: hermes-agent.nousresearch.com/docs/ +- @Teknium on X (Nous Research founder, posts on memory/skills architecture) + +## Key Architecture (from Leo's initial research) + +1. **4-layer memory system:** + - Prompt memory (MEMORY.md — always loaded, persistent identity) + - Session search (SQLite + FTS5 — conversation retrieval) + - Skills/procedural (reusable markdown procedures, auto-generated) + - Periodic nudge (autonomous memory evaluation) + +2. **7 pluggable memory providers:** Honcho, OpenViking (ByteDance), Mem0, Hindsight, Holographic, RetainDB, ByteRover + +3. **Skills = Taylor's instruction cards.** When agent encounters a task with 5+ tool calls, it autonomously writes a skill file. Uses agentskills.io open standard. Community skills via ClawHub/LobeHub. + +4. **Self-evolution repo (DSPy + GEPA):** Auto-submits improvements as PRs for human review + +5. **CamoFox:** Firefox fork with C++ fingerprint spoofing for web browsing + +6. **6 terminal backends:** local, Docker, SSH, Daytona, Singularity, Modal + +7. **Gateway layer:** Telegram, Discord, Slack, WhatsApp, Signal, Email + +8. **Release velocity:** 6 major releases in 22 days, 263 PRs merged in 6 days + +## Extraction Targets + +### NEW claims (ai-alignment): +1. Self-improving agent architectures converge on skill extraction as the primary learning mechanism (Hermes skills, Voyager skills, SWE-agent learned tools — all independently discovered "write a procedure when you solve something hard") +2. Agent self-evolution with human review gates is structurally equivalent to our governance model (DSPy + GEPA → auto-PR → human merge) +3. Memory architecture for persistent agents converges on 3+ layer separation (prompt/session/procedural/long-term) — Hermes, Letta, and our codex all arrived here independently + +### NEW claims (foundations/collective-intelligence): +4. Individual agent self-improvement (Hermes) is structurally different from collective knowledge accumulation (Teleo) — the former optimizes one agent's performance, the latter builds shared epistemic infrastructure +5. Pluggable memory providers suggest memory is infrastructure not feature — validates separation of knowledge store from agent runtime + +### ENRICHMENT candidates: +6. Enrich "Agentic Taylorism" claims — Hermes skills system is DIRECT evidence. Knowledge codification as markdown procedure files = Taylor's instruction cards. The agent writes the equivalent of a foreman's instruction card after completing a complex task. +7. Enrich collective-agent-core — Hermes architecture confirms harness > model (same model, different harness = different capability). Connects to Stanford Meta-Harness finding (6x performance gap from harness alone). + +## What They DON'T Do (matters for our positioning) + +- No epistemic quality layer (no confidence levels, no evidence requirements) +- No CI scoring or contribution attribution +- No evaluator role — self-improvement without external review +- No collective knowledge accumulation — individual optimization only +- No divergence tracking or structured disagreement +- No belief-claim cascade architecture + +This is the gap between agent improvement and collective intelligence. Hermes optimizes the individual; we're building the collective. + +## Pre-Screening Notes + +Check existing KB for overlap before extracting: +- `collective-agent-core.md` — harness architecture claims +- Agentic Taylorism claims in grand-strategy and ai-alignment +- Any existing Nous Research or Hermes claims (likely none) diff --git a/ops/queue.md b/ops/queue.md index 6564f730..1f3c078c 100644 --- a/ops/queue.md +++ b/ops/queue.md @@ -21,6 +21,7 @@ Outstanding work items visible to all agents. Everything here goes through eval | Identity reframe PRs need merging | review | medium | — | #149 Theseus, #153 Astra, #157 Rio, #158 Leo (needs rebase), #159 Vida. All have eval reviews. | | 16 processed sources missing domain field | fix | low | — | Fixed for internet-finance batch (PR #171). Audit remaining sources. | | Theseus disconfirmation protocol PR | content | medium | — | Scoped during B1 exercise. Theseus to propose. | +| Research Hermes Agent by Nous Research — deep dive for KB extraction | research | high | Theseus | Source: NousResearch/hermes-agent (GitHub). Research brief in `agents/theseus/musings/research-hermes-agent-nous.md`. **Extract:** (1) Skill extraction as convergent learning mechanism. (2) Self-evolution + human review gates = our governance model. (3) 3+ layer memory convergence. (4) Individual self-improvement ≠ collective knowledge accumulation. (5) Enrich Agentic Taylorism — skills = Taylor's instruction cards. Domains: ai-alignment + collective-intelligence. | ## Rules