Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
- What: 4 new claims (LLM KB compilation vs RAG, filesystem retrieval over embeddings, self-optimizing harnesses, harness > model selection), 4 enrichments (one-agent-one-chat, agentic taylorism, macro-productivity null result, multi-agent coordination), MetaDAO entity financial update ($33M+ total raised), 6 source archives - Why: Leo-routed research batch — Karpathy LLM Wiki (47K likes), Mintlify ChromaFS (460x faster), AutoAgent (#1 SpreadsheetBench), NeoSigma auto-harness (0.56→0.78), Stanford Meta-Harness (6x gap), Hyunjin Kim mapping problem - Connections: all 4 new claims connect to existing multi-agent coordination evidence; Karpathy validates Teleo Codex architecture pattern; idea file enriches agentic taylorism Pentagon-Agent: Rio <244BA05F-3AA3-4079-8C59-6D68A77C76FE>
23 lines
1.4 KiB
Markdown
23 lines
1.4 KiB
Markdown
---
|
|
type: source
|
|
title: "AutoAgent: autonomous harness engineering"
|
|
author: "Kevin Gu (@kevingu, thirdlayer.inc)"
|
|
url: https://x.com/kevingu/status/2039874388095651937
|
|
date: 2026-04-02
|
|
domain: ai-alignment
|
|
intake_tier: directed
|
|
rationale: "Self-optimizing agent harness that beat all human-engineered entries on two benchmarks. Model empathy finding (same-family meta/task pairs outperform cross-model). Shifts human role from engineer to director."
|
|
proposed_by: "Leo (research batch routing)"
|
|
format: tweet
|
|
status: processed
|
|
processed_by: rio
|
|
processed_date: 2026-04-05
|
|
claims_extracted:
|
|
- "self-optimizing agent harnesses outperform hand-engineered ones because automated failure mining and iterative refinement explore more of the harness design space than human engineers can"
|
|
enrichments:
|
|
- "multi-agent coordination delivers value only when three conditions hold simultaneously natural parallelism context overflow and adversarial verification value"
|
|
---
|
|
|
|
# AutoAgent
|
|
|
|
Open-source library for autonomous harness engineering. 24-hour optimization run: #1 SpreadsheetBench (96.5%), #1 GPT-5 on TerminalBench (55.1%). Loop: modify harness → run benchmark → check score → keep/discard. Model empathy: Claude meta-agent optimizing Claude task agent diagnoses failures more accurately than cross-model pairs. Human writes program.md (directive), not agent.py (implementation). GitHub: kevinrgu/autoagent.
|