leo: archive 19 tweet sources on AI agents, memory, and harnesses

- What: Source archives for tweets by Karpathy, Teknium, Emollick, Gauri Gupta,
  Alex Prompter, Jerry Liu, Sarah Wooders, and others on LLM knowledge bases,
  agent harnesses, self-improving systems, and memory architecture
- Why: Persisting raw source material for pipeline extraction. 4 sources already
  processed by Rio's batch (karpathy-gist, kevin-gu, mintlify, hyunjin-kim)
  were excluded as duplicates.
- Status: all unprocessed, ready for overnight extraction pipeline

Pentagon-Agent: Leo <D35C9237-A739-432E-A3DB-20D52D1577A9>
This commit is contained in:
m3taversal 2026-04-05 19:50:34 +01:00
parent 833f00a798
commit 00119feb9e
19 changed files with 465 additions and 0 deletions

View file

@ -0,0 +1,24 @@
---
type: source
title: "The Next Big Shift in AI Agents: Shared Context Graphs"
author: "Brana Rakic (@BranaRakic)"
url: "https://x.com/BranaRakic/status/2040159452431560995"
date: 2026-04-03
domain: ai-alignment
format: tweet
status: unprocessed
tags: [context-graphs, knowledge-base, agents, convergence]
---
## Content
Link to article: "The next big shift in AI agents: shared context graphs" - "Something interesting is converging. Karpathy is building personal knowledge bases with LLMs. Foundation Capital is writing about context graphs as the next..."
327 likes, 10 replies.
## Key Points
- Identifies convergence between Karpathy's personal knowledge bases and context graph concepts
- Shared context graphs proposed as the next major shift for AI agents
- Connects Foundation Capital's writing on context graphs to the broader trend
- Suggests a unified direction emerging from multiple independent developments

View file

@ -0,0 +1,23 @@
---
type: source
title: "NotebookLM Video on Karpathy Post"
author: "Emily (@IamEmily2050)"
url: "https://x.com/IamEmily2050/status/2040007450141593925"
date: 2026-04-03
domain: ai-alignment
format: tweet
status: unprocessed
tags: [notebooklm, karpathy-response, knowledge-base, video]
---
## Content
NotebookLM video overview on Andrej post.
1,173 likes, 22 replies. Video (~6 min) using NotebookLM to summarize Karpathy's knowledge base post.
## Key Points
- NotebookLM used to generate a video overview of Karpathy's LLM knowledge base post
- Demonstrates using one AI tool (NotebookLM) to summarize another AI workflow
- ~6 minute video summary

View file

@ -0,0 +1,24 @@
---
type: source
title: "Filesystems Replace RAG"
author: "Jerry Liu (@jerryjliu0)"
url: "https://x.com/jerryjliu0/status/2040154840228323468"
date: 2026-04-03
domain: ai-alignment
format: tweet
status: unprocessed
tags: [rag, filesystem, chromafs, mintlify, llamaindex, retrieval]
---
## Content
This is a cool article that shows how to *actually* make filesystems + grep replace a naive RAG implementation. Database + virtual filesystem abstraction + grep is all you need
780 likes, 28 replies. Includes image. Quotes Mintlify/ChromaFS article by Dens Sumesh. Jerry Liu is founder of LlamaIndex.
## Key Points
- Filesystems + grep can replace naive RAG implementations
- Database + virtual filesystem abstraction + grep is sufficient
- Endorsement from LlamaIndex founder of the filesystem-over-RAG approach
- References Mintlify/ChromaFS article as practical demonstration

View file

@ -0,0 +1,23 @@
---
type: source
title: "Towards Semantic Observability"
author: "Leonard Tang (@leonardtang_)"
url: "https://x.com/leonardtang_/status/2040122646197612557"
date: 2026-04-03
domain: ai-alignment
format: tweet
status: unprocessed
tags: [observability, monitoring, ai-systems, infrastructure]
---
## Content
Link to article: "Towards Semantic Observability" - discusses how traditional observability relies on knowing failure behaviors in advance.
353 likes, 10 replies.
## Key Points
- Traditional observability assumes you know failure behaviors in advance
- Proposes semantic observability as an alternative approach for AI systems
- Addresses the challenge of monitoring systems with unpredictable failure modes

View file

@ -0,0 +1,24 @@
---
type: source
title: "LLM Knowledge Base System Diagram"
author: "omarsar0 (@omarsar0)"
url: "https://x.com/omarsar0/status/2040099881008652634"
date: 2026-04-03
domain: ai-alignment
format: tweet
status: unprocessed
tags: [llm, knowledge-base, diagram, karpathy-response, visualization]
---
## Content
Diagram of the LLM Knowledge Base system. Feed this to your favorite agent and get your own LLM knowledge base going.
1,624 likes, 49 replies. Contains diagram image of Karpathy's 3-layer system.
## Key Points
- Provides a diagram of Karpathy's LLM Knowledge Base system architecture
- 3-layer system design visualized
- Designed to be fed to an agent to bootstrap your own knowledge base
- Practical starter resource for implementing the pattern

View file

@ -0,0 +1,24 @@
---
type: source
title: "Become a Generalist"
author: "oprydai (@oprydai)"
url: "https://x.com/oprydai/status/2040130116022661243"
date: 2026-04-03
domain: ai-alignment
format: tweet
status: unprocessed
tags: [generalism, cross-domain, innovation, patterns]
---
## Content
become a generalist. specialization makes you efficient. generalization makes you dangerous. what it actually means: learn across domains -- math, physics, software, economics, biology. patterns repeat across fields. connect ideas -- innovation happens at the intersection
5,115 likes, 210 replies. Includes attached image.
## Key Points
- Specialization makes you efficient but generalization makes you dangerous
- Learning across domains (math, physics, software, economics, biology) reveals repeating patterns
- Innovation happens at the intersection of ideas from different fields
- Cross-domain pattern recognition is a key competitive advantage

View file

@ -0,0 +1,24 @@
---
type: source
title: "Why Memory Isn't a Plugin (It's the Harness)"
author: "Sarah Wooders (@sarahwooders)"
url: "https://x.com/sarahwooders/status/2040121230473457921"
date: 2026-04-03
domain: ai-alignment
format: tweet
status: unprocessed
tags: [memory, agent-harness, letta-ai, memgpt]
---
## Content
Link to article: "Why memory isn't a plugin (it's the harness)" - discusses MemGPT/Letta AI's memory architecture. Argues memory should be the harness, not a plugin bolted on. Associated with Letta AI.
316 likes, 10 replies.
## Key Points
- Memory should be the harness, not a plugin bolted onto an agent
- Discusses MemGPT/Letta AI's memory architecture
- Challenges the common pattern of treating memory as an add-on component
- Positions memory as fundamental infrastructure rather than optional feature

View file

@ -0,0 +1,24 @@
---
type: source
title: "Hermes Agent v0.7 Memory Deep Dive"
author: "Teknium (@Teknium)"
url: "https://x.com/Teknium/status/2040151297991770435"
date: 2026-04-03
domain: ai-alignment
format: tweet
status: unprocessed
tags: [hermes-agent, nous-research, memory, interfaces, architecture]
---
## Content
Deeper dive into some of the updates in v0.7. Memory: We have begun transitioning each of the systems in Hermes Agent to work through defined interfaces so that the core code is more maintainable, and more providers for everything can be supported. We started with memory:
375 likes, 36 replies. Includes attached image of memory architecture. Quote of NousResearch announcement.
## Key Points
- Hermes Agent v0.7 transitions systems to work through defined interfaces
- Interface-based architecture improves maintainability and extensibility
- Memory system was the first to be refactored to this interface pattern
- Enables support for multiple providers per system component

View file

@ -0,0 +1,25 @@
---
type: source
title: "Stanford Meta-Harness: Biggest Performance Gap Is the Harness"
author: "alex_prompter (@alex_prompter)"
url: "https://x.com/alex_prompter/status/2040378405322113442"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [harness, meta-harness, stanford, agent-optimization, benchmark]
---
## Content
Holy shit. Stanford just showed that the biggest performance gap in AI systems isn't the model it's the harness. The code wrapping the model. And they built a system that writes better harnesses automatically than humans can by hand. +7.7 points. 4x fewer tokens. #1 ranking
613 likes, 32 replies. Contains research visualization image.
## Key Points
- Stanford research shows the harness (code wrapping the model) matters more than the model itself
- Built a system that automatically writes better harnesses than human-crafted ones
- Achieved +7.7 point improvement with 4x fewer tokens
- Reached #1 ranking on benchmark
- Key implication: optimizing the harness is higher leverage than optimizing the model

View file

@ -0,0 +1,25 @@
---
type: source
title: "515 Startup Field Experiment on AI Adoption"
author: "Ethan Mollick (@emollick)"
url: "https://x.com/emollick/status/2040436307176898897"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [ai-adoption, startups, field-experiment, productivity, mapping-problem]
---
## Content
Big deal paper here: field experiment on 515 startups, half shown case studies of how startups are successfully using AI. Those firms used AI 44% more, had 1.9x higher revenue, needed 39% less capital: 1) AI accelerates businesses 2) The challenge is understanding how to use it
995 likes. Includes 2 images. Quotes Hyunjin Kim's paper on AI's "mapping problem" in firms.
## Key Points
- Field experiment on 515 startups showed significant AI adoption effects
- Firms shown AI case studies used AI 44% more than control group
- Treatment group had 1.9x higher revenue and needed 39% less capital
- The main challenge is not AI capability but understanding how to use it
- References the "mapping problem" -- discovering where AI creates value

View file

@ -0,0 +1,29 @@
---
type: source
title: "auto-harness: Self-Improving Agentic Systems with Auto-Evals"
author: "Gauri Gupta (@gauri__gupta)"
url: "https://x.com/gauri__gupta/status/2040251309782409489"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [auto-harness, self-improving, auto-evals, open-source, agent-optimization]
---
## Content
Releasing auto-harness: an open source library for our self improving agentic systems with auto-evals. We got a lot of responses from people wanting to try the self-improving loop on their own agent. So we open-sourced our setup. Connect your agent and let it cook over the...
371 likes, 11 replies. Links to article about self-improving agentic systems.
Additional tweet (https://x.com/gauri__gupta/status/2040251170099524025):
Link to article: "auto-harness: Self improving agentic systems with auto-evals (open-sourced!)" - "a self-improving loop that finds your agent's failures, turns them into evals, and fixes them."
1,100 likes, 15 replies.
## Key Points
- auto-harness is an open-source library for self-improving agentic systems
- Implements a self-improving loop: find failures, turn them into evals, fix them
- Open-sourced in response to community demand
- Connect your own agent to the self-improving loop
- Automatic evaluation generation from observed failures

View file

@ -0,0 +1,25 @@
---
type: source
title: "6 Components of Coding Agents"
author: "Hesamation (@Hesamation)"
url: "https://x.com/Hesamation/status/2040453130324709805"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [coding-agents, harness, claude-code, components, architecture]
---
## Content
this is a great article if you want to understand Claude Code or Codex and the main components of a coding agent: 'harness is often more important than the model'. LLM -> agent -> agent harness -> coding harness. there are 6 critical components: 1. repo context: git, readme, ...
279 likes, 15 replies. Quote of Sebastian Raschka's article on coding agent components.
## Key Points
- Harness is often more important than the model in coding agents
- Layered architecture: LLM -> agent -> agent harness -> coding harness
- 6 critical components identified, starting with repo context (git, readme)
- Applicable to understanding Claude Code and Codex architectures
- References Sebastian Raschka's detailed article on the topic

View file

@ -0,0 +1,23 @@
---
type: source
title: "Karpathy KB Architecture Visualization"
author: "Himanshu (@himanshustwts)"
url: "https://x.com/himanshustwts/status/2040477663387893931"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [llm, knowledge-base, architecture, visualization, karpathy-response]
---
## Content
this is beautiful. basically a pattern for building personal knowledge bases using LLMs. and here is the architecture visualization of what karpathy says as 'idea file'. i think this is quite hackable / experimental and numerous things can be explored from here
806 likes, 14 replies. Includes attached image visualization of the architecture.
## Key Points
- Provides an architecture visualization of Karpathy's LLM knowledge base pattern
- Frames the pattern as hackable and experimental
- Suggests numerous directions for exploration from this base pattern

View file

@ -0,0 +1,24 @@
---
type: source
title: "EPUB to TXT via Agents"
author: "Andrej Karpathy (@karpathy)"
url: "https://x.com/karpathy/status/2040451573881737480"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [llm, agents, epub, conversion, karpathy]
---
## Content
@trainable_nick The best epub to txt converter I found is just asking your favorite agent to do it. Epubs can be very diverse, the agent just goes in, figures it out, creates the output markdown and ensures it looks good works great.
976 likes, 44 replies. Reply to trainable_nick about EPUB conversion tools.
## Key Points
- LLM agents can serve as the best EPUB to text converters
- Agents handle the diversity of EPUB formats by figuring out structure dynamically
- Agents can ensure output quality by reviewing their own work
- Practical example of agents replacing specialized tooling

View file

@ -0,0 +1,24 @@
---
type: source
title: "Idea Files for the LLM Era"
author: "Andrej Karpathy (@karpathy)"
url: "https://x.com/karpathy/status/2040470801506541998"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [llm, agents, idea-file, knowledge-sharing, karpathy]
---
## Content
Wow, this tweet went very viral! I wanted share a possibly slightly improved version of the tweet in an 'idea file'. The idea of the idea file is that in this era of LLM agents, there is less of a point/need of sharing the specific code/app, you just share the idea, then the other person's agent customizes & builds it.
21,135 likes, 761 replies. Links to GitHub Gist "llm-wiki".
## Key Points
- In the LLM agent era, sharing ideas is more valuable than sharing specific code
- "Idea files" allow others' agents to customize and build implementations
- Follow-up to the viral LLM Knowledge Bases post
- Links to a GitHub Gist called "llm-wiki" as an example idea file

View file

@ -0,0 +1,28 @@
---
type: source
title: "Claude Code Skills Guide"
author: "nyk (@nyk_builderz)"
url: "https://x.com/nyk_builderz/status/2040391725391516065"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [claude-code, skills, agent-harness, prompt-engineering]
---
## Content
If Claude keeps repeating the same mistakes, you don't need a longer prompt - you need a skill. I wrote a practical guide to building Claude Code skills that auto-invoke when relevant: SKILL.md structure, trigger design, allowed-tools safety, templates/examples
42 likes, 4 replies. Links to article "Build Claude Code Skills: The full guide".
Additional tweet (https://x.com/nyk_builderz/status/2040338207188062270):
"Build Claude Code Skills: The full guide" - "Most Claude Code skill guides overcomplicate something that's actually simple. Here's the version that actually works."
100 likes, 4 replies.
## Key Points
- Claude Code skills auto-invoke when relevant, replacing longer prompts
- Guide covers SKILL.md structure, trigger design, and allowed-tools safety
- Skills address repeating mistakes by encoding reusable patterns
- Practical templates and examples provided

View file

@ -0,0 +1,24 @@
---
type: source
title: "Hermes Agent v0.7 Pluggable Memory"
author: "sudoingX (@sudoingX)"
url: "https://x.com/sudoingX/status/2040408975246856569"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [hermes-agent, nous-research, memory, pluggable-architecture]
---
## Content
holy shit hermes agent v0.7.0 just dropped and your memory is now fully pluggable. 7 providers out of the box from cloud to local sqlite. don't like any of them? build your own and plug it in. credential pools. multiple API keys per provider with automatic rotation. key gets...
166 likes, 9 replies. Quote of Teknium's post about Hermes Agent v0.7.
## Key Points
- Hermes Agent v0.7.0 introduces fully pluggable memory with 7 providers
- Memory providers range from cloud to local SQLite
- Custom memory providers can be built and plugged in
- Credential pools with automatic API key rotation added

View file

@ -0,0 +1,24 @@
---
type: source
title: "EPUB to Markdown Tool"
author: "trainable_nick (@trainable_nick)"
url: "https://x.com/trainable_nick/status/2040448094060343337"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [epub, markdown, vibe-coding, knowledge-base, tool]
---
## Content
As I pulled on the thread from Karpathy's post, I realized the existing EPUB to TXT tools were still too ugly and clunky for turning DRM-free books into clean markdown. So I made my own. I've only been vibe coding for a few months, and this is my first App Store Connect
239 likes, 11 replies. Includes image. Quote of Karpathy's KB post.
## Key Points
- Existing EPUB to TXT tools were insufficient for clean markdown output
- Built a new tool specifically for converting DRM-free books to clean markdown
- Inspired directly by Karpathy's LLM knowledge base workflow
- Creator's first App Store Connect submission, built via vibe coding

View file

@ -0,0 +1,24 @@
---
type: source
title: "Karpathy's LLM Wiki Pattern"
author: "Yuchen J (@Yuchenj_UW)"
url: "https://x.com/Yuchenj_UW/status/2040482771576197377"
date: 2026-04-04
domain: ai-alignment
format: tweet
status: unprocessed
tags: [llm, knowledge-base, wiki, karpathy-response]
---
## Content
Karpathy's 'LLM Wiki' pattern: stop using LLMs as search engines over your docs. Use them as tireless knowledge engineers who compile, cross-reference, and maintain a living wiki. Humans curate and think.
1,352 likes, 45 replies. Includes a diagram generated by Claude agent.
## Key Points
- Reframes LLM usage from search engine to knowledge engineer
- LLMs should compile, cross-reference, and maintain living wikis
- Humans retain the curation and thinking roles
- Distillation of Karpathy's LLM Knowledge Base workflow